## Lists and List Comprehensions
1. In Kapulet, what does `[ [x, 2*x] | x from [1, 2, 3] ]` evaluate to?
> [[1,2], [2,4], [3,6]]
2. What does `[ 10*x + y | y from [4], x from [1, 2, 3] ]` evalaute to?
> [14, 24, 34]
3. Using either Kapulet's or Haskell's list comprehension syntax, write an expression that transforms `[3, 1, 0, 2]` into `[3, 3, 3, 1, 2, 2]`. [[Here is a hint|assignment3 hint1]], if you need it.
> letrec
> dup (n, x) = case n of
> 0 then [];
> _ then x & dup (n-1, x)
> end;
> # you can assume join is already defined, but we'll redefine it here anyway
> join xss = fold_right ((&&), []) xss # we help ourselves to &&/append
> in join [dup(x, x) | x from [3, 1, 0, 2]]
4. Last week you defined `head` in terms of `fold_right`. Your solution should be straightforwardly translatable into one that uses our proposed right-fold encoding of lists in the Lambda Calculus. Now define `empty?` in the Lambda Calculus. (It should require only small modifications to your solution for `head`.)
> let empty? = \xs. xs (\x z. false) true in
> ...
5. If we encode lists in terms of their *left*-folds, instead, `[a, b, c]` would be encoded as `\f z. f (f (f z a) b) c`. The empty list `[]` would still be encoded as `\f z. z`. What should `cons` be, for this encoding?
> let left_cons = \x xs. \f z. xs f (f z x) in
> ...
6. Continuing to encode lists in terms of their left-folds, what should `last` be, where `last [a, b, c]` should result in `c`. Let `last []` result in whatever `err` is bound to.
> let left_last = \xs. xs (\z x. x) err in
> ...
7. Continuing to encode lists in terms of their left-folds, how should we write `head`? This is challenging.
> The [[hint/solution|assignment3 hint2]] gave this answer:
> let m = \a. \n b. n a in
> let f = \w b. w m b in
> let left_head = \xs. (xs f I) I err in
> ...
> Here's another, more straightforward answer. We [[observed before|/topics/week2_encodings#flipped-cons]] that left-folding the `cons` function over a list reverses it. Hence, it is easy to reverse a list defined in terms of its left-fold:
> let left_empty = \f z. z in ; this the same as for the right-fold encoding of lists
> let flipped_left_cons = \xs x. \f z. xs f (f z x) in ; left-folding supplies arguments in a different order than cons expects
> let left_reverse = \xs. xs flipped_left_cons left_empty in
> ...
> Then we can get the list's head by reversing it and getting the last member of the result, which we saw in the previous problem is easy:
> let left_head = \xs. left_last (left_reverse xs) in
> ...
> I'm not sure which of the two solutions presented here is better. The one given in the hint traverses the list only once; whereas the one gotten by reversing the list and getting the last member of the result traverses the list twice. But the former strategy does more complicated stuff at each step of the traversal (both conceptually and more applications), so in the end it might be computationally "cheaper" to use the latter strategy.
8. Suppose you have two lists of integers, `left` and `right`. You want to determine whether those lists are equal, that is, whether they have all the same members in the same order. How would you implement such a list comparison? You can write it in Scheme or Kapulet using `letrec`, or if you want more of a challenge, in the Lambda Calculus using your preferred encoding for lists. If you write it in Scheme, don't rely on applying the built-in comparison operator `equal?` to the lists themselves. (Nor on the operator `eqv?`, which might not do what you expect.) You can however rely on the comparison operator `=` which accepts only number arguments. If you write it in the Lambda Calculus, you can use your implementation of `leq?`, requested below, to write an equality operator for Church-encoded numbers. [[Here is a hint|assignment3 hint3]], if you need it.
(The function you're trying to define here is like `eqlist?` in Chapter 5 of *The Little Schemer*, though you are only concerned with lists of numbers, whereas the function from *The Little Schemer* also works on lists containing symbolic atoms --- and in the final version from that Chapter, also on lists that contain other, embedded lists.)
> In Kapulet:
> letrec
> list_eq? (left, right) = case (left, right) of
> ([], []) then 'true;
> ([], _) then 'false;
> (_, []) then 'false;
> (x&xs, y&ys) when x == y then list_eq? (xs, ys);
> (_, _) then 'false
> end
> in list_eq?
> In Scheme:
> (define list_eq? (lambda (left right)
> (cond
> ((null? left) (null? right))
> ((null? right) #f)
> ((= (car left) (car right)) (list_eq? (cdr left) (cdr right)))
> (else #f))))
> In the Lambda Calculus, helping ourselves to definitions of `leq?`, `0`, `succ`, `head`, `tail`, and `reverse`:
> let true = \y n. y in
> let and = \p q. p q p in
> let pair = \a b. \f. f a b in
> let fst = \x y. x in
> let num_eq? = \l r. and (leq? l r) (leq? r l) in
> let length = \xs. xs (\x z. succ z) 0 in
> let check = \x p. p (\b ys. pair (and b (num_eq? x (head ys))) (tail ys)) in
> let list_eq? = \xs ys. and (num_eq? (length xs) (length ys)) ; this step could be skipped
> ; if you were prepared to assume the lists always have the same length
> (xs check (pair true (reverse ys)) fst) in
> ...
## Numbers
9. Recall our proposed encoding for the numbers, called "Church's encoding". As we explained last week, it's similar to our proposed encoding of lists in terms of their folds. Give a Lambda Calculus definition of `zero?` for numbers so encoded. (It should require only small modifications to your solution for `empty?`, above.)
> let zero? = \n. n (\z. false) true in
> ...
10. In last week's homework, you gave a Lambda Calculus definition of `succ` for Church-encoded numbers. Can you now define `pred`? Let `pred 0` result in whatever `err` is bound to. This is challenging. For some time theorists weren't sure it could be done. (Here is [some interesting commentary](http://okmij.org/ftp/Computation/lambda-calc.html#predecessor).) However, in this week's notes we examined one strategy for defining `tail` for our chosen encodings of lists, and given the similarities we explained between lists and numbers, perhaps that will give you some guidance in defining `pred` for numbers.
> Here is the solution we were guiding you towards, due to Paul Bernays:
> ; helping ourselves to 0 and succ
> let pair = \a b. \f. f a b in
> let snd = \x y. y in
> let shift = \p. p (\x y. pair (succ x) x) in
> let pred = \n. n shift (pair 0 err) snd in
> ...
> Here is another solution, due to Martin Bunder and F. Urbanek:
> let box = \a. \v. v a in
> let pred = \n. \s z. n (\b. box (b s)) (K z) I in
> ...
> That can be hard to wrap your head around. If you trace this expression through, you'll see that:
> * when `n == 0`, it reduces to `\s z. (K z) I`, or `\s z. z`
> * when `n == 1`, it reduces to `\s z. (box z) I`, or `\s z. z`
> * when `n == 2`, it reduces to `\s z. (box (s z)) I`, or `\s z. s z`
> * when `n == 3`, it reduces to `\s z. (box (s (s z))) I`, or `\s z. s (s z)`
> `box a` is `\v. v a`; this stands to `pair a b` as one stands to two. Since boxes (like pairs) are really just functions, the technique used in this definition of `pred` is akin to that used in [[the hint for last week's `reverse`|assignment2_answers#cps-reverse]].
(Want a further challenge? Define `map2` in the Lambda Calculus, using our right-fold encoding for lists, where `map2 g [a, b, c] [d, e, f]` should evaluate to `[g a d, g b e, g c f]`. Doing this will require drawing on a number of different tools we've developed, including that same strategy for defining `tail`. Purely extra credit.)
> let next = \g. \x p. p (\zs ys. pair (cons (g x (head ys)) zs) (tail ys)) in
> let finish = \zs ys. (empty? ys) zs
> ; else ys are too long
> err_different_lengths in
> let map2 = \g. \xs ys. (leq? (length xs) (length ys))
> (xs (next g) (pair empty (reverse ys)) finish)
> ; else ys are too short
> err_different_lengths in
> ...
11. Define `leq?` for numbers (that is, ≤) in the Lambda Calculus. Here is the expected behavior,
where `one` abbreviates `succ zero`, and `two` abbreviates `succ (succ zero)`.
leq? zero zero ~~> true
leq? zero one ~~> true
leq? zero two ~~> true
leq? one zero ~~> false
leq? one one ~~> true
leq? one two ~~> true
leq? two zero ~~> false
leq? two one ~~> false
leq? two two ~~> true
...
You'll need to make use of the predecessor function, but it's not essential to understanding this problem that you have successfully implemented it yet. You can treat it as a black box.
> It's easiest to do this if you define `pred` so that `pred 0 ~~> 0`:
> let pred = \n. n shift (pair 0 0) snd in
> let sub = \l r. r pred l in
> let leq? = \l r. zero? (sub l r) in
> ...
> Here is another solution. Jim crafted this particular implementation, but like a great deal of the CS knowledge he's gained over the past eight years, Oleg Kiselyov pointed the way.
> let leq? = (\base build consume. \l r. r consume (l build base) fst)
> ; where base is
> (pair true junk)
> ; and build is
> (\p. pair false p)
> ; and consume is
> (\p. p fst p (p snd)) in
> ...
> That can be generalized to give this nice implementation of `num_eq?`:
> ; `num_cmp x y lt eq gt` returns lt when xy
> let num_cmp = (\base build consume. \l r. r consume (l build base) fst)
> ; where base is
> (pair (\a b c. b) (K (\a b c. a)))
> ; and build is
> (\p. pair (\a b c. c) p)
> ; and consume is
> (\p. p fst p (p snd) (p snd)) in
> let num_eq? = \x y. num_cmp x y false true false in
> ...
## Combinatory Logic
Reduce the following forms, if possible:
`Kxy ~~> x`

`KKxy ~~> Ky`

`KKKxy ~~> Kxy ~~> x`

`SKKxy ~~> Kx(Kx)y ~~> xy`

`SIII ~~> II(II) ~~> II ~~> I`

`SII(SII) ~~> I(SII)(I(SII)) ~~> SII(SII) ~~> ...`

; the reduction never terminates

18. Give Combinatory Logic combinators (that is, expressed in terms of `S`, `K`, and `I`) that behave like our boolean functions. Provide combinators for `true`, `false`, `neg`, `and`, `or`, and `xor`.
> As Dan pointed out in the Wednesday session, this problem is easy to solve if you first come up with a Combinatory translation of the schema `\p. if p then M else N`:
> [\p. p M N] =
> @p. p [M] [N] =
> S (@p. p [M]) (@p [N]) =
> S (S (@p p) (@p [M])) (@p [N]) =
> S (S I (@p [M])) (@p [N]) =
> S (S I (K [M])) (K [N])
> The last step assumes that `p` does not occur free in the expressions `M` and `N`.
> Next, we translate `true ≡ \y n. y` as `K`, and `false ≡ \y n. n` as `KI`. Then we rely on the following equivalences:
> neg ≡ \p. if p then false else true
> and ≡ \p. if p then I else K false
> ; then: and p q <~~> if p then I q else K false q ~~> if p then q else false
> or ≡ \p. if p then K true else I
> ; then: or p q <~~> if p then K true q else I q ~~> if p then true else q
> xor ≡ \p. if p then neg else I
> ; then: xor p q <~~> if p then neg q else I q ~~> if p then neg q else q
> Then we just make use of the schema derived above to complete the translations. For example, `and ≡ \p. if p then I else K false` becomes `S (SI (K `**I**)) (K **(K false)**) ≡ S (SI (K **I**)) (K **(K (KI))**)

.
> In the case of `neg`, this produces `S (SI (K `**false**)) (K **true**) ≡ S (SI (K **(KI)**)) (K **K**)

. A more compact result here can be obtained here by observing that for boolean arguments, `neg` can also be defined as `\p y n. p n y`, which is just the `C` combinator.
Using the mapping specified in this week's notes, translate the following lambda terms into combinatory logic:
`[\x x] = @x x = I`

`[\x y. x] = @x [\y. x] = @x. (@y x) = @x (Kx) = S (@x K) (@x x) = S (KK) I`

; in general expressions of this form `S(K`*M*)I

will behave just like *M*

for any expression *M*

`[\x y. y] = @x [\y. y] = @x (@y y) = @x I = KI`

`[\x y. y x] = @x [\y. y x] = @x (@y (y x)) = @x (S (@y y) (@y x)) = @x (S I (Kx)) = S (@x (SI)) (@x (Kx)) = S (K(SI)) K`

; this is the **T** combinator
`[\x. x x] = @x (x x) = S (@x x) (@x x) = SII`

`[\x y z. x (y z)] =`

@x (@y (@z (x (y z)))) =

@x (@y (S (@z x) (@z (y z)))) =

@x (@y (S (Kx) (@z (y z)))) =

@x (@y (S (Kx) y)) =

@x (S (@y (S (Kx))) (@y y)) =

@x (S (K (S (Kx))) (@y y)) =

@x (S (K (S (Kx))) I) =

S (@x (S (K (S (Kx))))) (@x I) =

S (S (@x S) (@x (K (S (Kx))))) (@x I) =

S (S (KS) (@x (K (S (Kx))))) (@x I) =

S (S (KS) (S (@x K) (@x (S (Kx))))) (@x I) =

S (S (KS) (S (KK) (@x (S (Kx))))) (@x I) =

S (S (KS) (S (KK) (S (@x S) (@x (Kx))))) (@x I) =

S (S (KS) (S (KK) (S (KS) (@x (Kx))))) (@x I) =

S (S (KS) (S (KK) (S (KS) K))) (@x I) =

S (S (KS) (S (KK) (S (KS) K))) (KI)

; this is the **B** combinator, which can also be expressed more concisely as `S (K S) K`

25. For each of the above translations, how many `I`s are there? Give a rule for describing what each `I` corresponds to in the original lambda term.
This generalization depends on you omitting the translation rule:
6. @a(Xa) = X if a is not in X
> With that shortcut rule omitted, then there turn out to be one `I` in the result corresponding to each occurrence of a bound variable in the original term.
Evaluation strategies in Combinatory Logic
------------------------------------------
26. Find a term in CL that behaves like Omega does in the Lambda
Calculus. Call it `Skomega`.
> SII(SII)
27. Are there evaluation strategies in CL corresponding to leftmost
reduction and rightmost reduction in the Lambda Calculus?
What counts as a redex in CL?
> A redex is any `I` with one argument, or `K` with two arguments, or `S` with three arguments.
> Leftmost reduction corresponds to applying the rules for an `I` or `K` or `S` in head position *before* reducing any of their arguments; rightmost reduction corresponds to reducing the arguments first.
28. Consider the CL term `K I Skomega`. Does leftmost (alternatively,
rightmost) evaluation give results similar to the behavior of `K I
Omega` in the Lambda Calculus, or different? What features of the
Lambda Calculus and CL determine this answer?
> Leftmost reduction terminates with `I`; rightmost reduction never terminates. This is because "leftmost reduction" corresponds to "head-redex first", which gives the head combinator the opportunity to ignore some of its (in this case, unreducible) arguments without trying to reduce them. If the arguments are reduced first, the head combinator loses that opportunity.
29. What should count as a thunk in CL? What is the equivalent
constraint in CL to forbidding evaluation inside of a lambda abstract?
> Recall that a thunk is a function of the form `\(). EXPRESSION`, where `EXPRESSION` won't be evaluated until the function is applied (to the uninformative argument `()`). We don't really have `()` in the Lambda Calculus. The closest approximation would be `I`, since a triple is `\f. f a b c`, a pair is `\f. f a b`, and following the same pattern a one-tuple would be `\f. f a` (not the same as `a`), and the zero-tuple would be `\f. f` or `I`. But really in the untyped Lambda Calculus, it doesn't matter whether you supply `I` to an abstract that's going to ignore its argument, or any other argument. Instead, you could get the same effect as a thunk by just using `\x. EXPRESSION`, where `x` is a variable that doesn't occur free in `EXPRESSION`.
> The way to do the same thing in Combinatory Logic is `K (EXPRESSION)`. The constraint we need is that arguments to `S` and `K` are never reduced until there are enough of them to apply the rule for `S` and `K`. (If we furthermore insist on leftmost reduction, then expressions will *never* be reduced when in the position of an argument to another combinator. Only the reduction rules for the outermost head combinator are ever followed.)
More Lambda Practice
--------------------
Reduce to beta-normal forms:
`(\x. x (\y. y x)) (v w) ~~> v w (\y. y (v w))`

`(\x. x (\x. y x)) (v w) ~~> v w (\x. y x)`

`(\x. x (\y. y x)) (v x) ~~> v x (\y. y (v x))`

`(\x. x (\y. y x)) (v y) ~~> v y (\u. u (v y))`

`(\x y. x y y) u v ~~> u v v`

`(\x y. y x) (u v) z w ~~> z (u v) w`

`(\x y. x) (\u u) ~~> \y u. u`

, also known as `false`

`(\x y z. x z (y z)) (\u v. u) ~~> \y z. (\u v. u) z (y z) ~~> \y z. z`

, also known as `false`