# Anonymous functions and closures

Anonymous functions and closures are an important part of the Julia syntax.

For example, a simple example of the frequent use of anonymous functions is on calls to the findfirst function, which returns the element of an array which first matches a condition:

julia> x = [ 0, π/4, π/2, π ]
4-element Array{Float64,1}:
0.0
0.7853981633974483
1.5707963267948966
3.141592653589793

julia> findfirst( x -> sin(x) > 0.5, x )
2

The x -> sin(x) > 0.5 is an anonymous function, which one can read as "given x return sin(x) > 0.5", which in this case can be true or false.

## Basic syntax and closures

Consider the following function:

a = 5
f(x,a) = a*x

This function is also a "closure", because it "closes over" the variable a. Be careful, if written in global scope, this a is type-unstable, unless if declared const.

We can define an anonymous function which, given x, returns the result of f(x,a):

x -> f(x,a) 

The anonymous function can be bound to a label name, as any other value:

g = x -> f(x,a)

and g will be a function that, given x, returns f(x,a). This definition is similar to

g(x) = f(x,a)

Except that in the latter case the label g is bound to the function in definitive:

julia> g(x) = f(x,a)
g (generic function with 1 method)

julia> g = 2
ERROR: invalid redefinition of constant g

While in the former case g can be redefined at will, as it is only a label bound to the address of an anonymous function:

julia> g = x -> f(x,a)
#1 (generic function with 1 method)

julia> g = 2
2

Anonymous functions for functions with multiple parameters, and closing over multiple values can be defined likewise:

(x,y) -> f(x,y,a,b,c)

## Use case

As exemplified for the findfirst function, the most important use for anonymous functions is passing functions as arguments to other functions, in particular when the function needs additional data to be evaluated. Consider the following toy example which reproduces a common scenario in which anonymous functions appear:

Let us suppose we have a package which implements a solver for an optimization problem. That is, the solver receives a function and an initial point, and returns the minimizer of that function. A typical call for such a solver would be

x = solver(f,x0)

where f is the function to be minimized and x0 the initial guess. Within the solver code you will find calls to f of the form f(x).

If f is a function like f(x) = x^2 + 2x - 3, we can just define the function and call the solver with f.

However, what if f depends on data? For example, if f was defined as:

f(x,a,b,c) = a*x^2 + b*x + c

The solver does not explicitly support calls to f with general arguments, because that would be cumbersome. How can we then call the solver with a function which will be interpreted by the solver as f(x) but uses the parameters a, b, and c? That is what the anonymous functions solve here.

## Example

For example, lets create a "solver" that only evaluates the function f and returns:

julia> function solver(f,x)
y = f(x)
return y
end
solver (generic function with 1 method)

Then, let us define a function that depends on three constant parameters besides that the variable x:

julia> const a, b, c = 1, 2, 3;

julia> g(x) = a*x^2 + b*x + c

We can do, now:

julia> x = 5.;

julia> solver(g,x)
38.0

Alternatively, one could use an anonymous function:

julia> solver(x -> a*x^2 + b*x + c,x)
38.0

We could, also, use g = x -> a*x^2 + b*x + c, but this option will be less preferred as it is just a wacky way to write g(x) = a*x^2 + b*x +c.

## Scope of variables

In the examples of the previous section, the parameters a, b, and c of the function g were defined in the global scope will cause type-instability problems. In the definition of the function as

g(x) = a*x^2 + b*x + c

it is quite evident, as g does not receive the parameters as arguments (the purpose of the definition of g was that one) and thus g is using those parameters from the global scope.

Using the anonymous functions the scope of the parameters is the same, even if this is less evident. In

solver(x -> a*x^2 + b*x + c, x) 

the anonymous function x -> a*x^2 + b*x + c was parsed at the calling scope, not at the scope of the solver. Therefore, except for not having a name, it behaves exactly as the former g(x), thus the parameters are non-constant globals and will introduce type instabilities and performance penalties.

The parameters of the anonymous function have to be, therefore, constant for the code to be type-stable and performant. This is why we declared them as constants,

julia> const a, b, c = 1, 2, 3;

Alternatively (and better, because it is more flexible), we can wrap the call to the solver in a function that receives the data as parameters, in which case the scope of the data will be the scope of the calling function of the solver, not the global scope:

julia> a, b, c = 1, 2, 3; # not necessarily constant

julia> function h(x,a,b,c)
return solver(x -> a*x^2 + b*x + c,x)
end
h (generic function with 1 method)

julia> x = 5.0;

julia> h(x,a,b,c)
38.0


(one could have written h(x,a,b,c) = solver(x->a*x^2 + b*x + c,x), but the syntax above is more explicit in the fact that h has its own scope of variables).

# Take away

Passing functions as argument to other functions, in particular if the evaluation of data is required, is practical with the use of anonymous functions and closures. This occurs very frequently in the context of calls to solvers, but very frequently also in the Julia base language, for example in the search and sorting, functions, among many others. One has to keep in mind that the closures are parsed at the calling scope, such that critical code for performance must always be wrapped into functions, to guarantee the constant types of the parameters. Let us just reinforce this point with an example.

# Example of type-instability

Let us write a function that operates on a vector, returning a some "potential energy" associated to some computation on pair of elements of the vector:

julia> function u(f,x)
u = 0.
for i in 1:length(x)-1
for j in i+1:length(x)
u += f(x[i],x[j])
end
end
u
end
u (generic function with 1 method)


Let use define f as

julia> f(x,y,a,b) = a*x - b*y
f (generic function with 1 method)


And we will use an anonymous function to pass the function f to u:

julia> a = 5; b = 7;

julia> u( (x,y) -> f(x,y,a,b), x )
-567899.3283195692


An analysis of the stability of the types will indicate type-instabilities, associated with the global scope of the parameters a and b:

julia> @code_warntype u( (x,y) -> f(x,y,a,b), x )
Variables
#self#::Core.Compiler.Const(u, false)
...
x::Array{Float64,1}
u::Any
@_5::Union{Nothing, Tuple{Int64,Int64}}
...
Body::Any
1 ─       (u = 0.0)
...
│   %22 = u::Any
│   %23 = Base.getindex(x, i)::Float64
│   %24 = Base.getindex(x, j)::Float64
│   %25 = (f)(%23, %24)::Any
│         (u = %22 + %25)
...


If, alternatively, one defines an enclosing function for the call to the energy function,

julia> h(x,a,b) =  u( (x,y) -> f(x,y,a,b), x )
h (generic function with 1 method)


The type-instabilities are resolved, because the scope in which the anonymous functions is parsed is now the local scope of h and, thus, all variables have constant types:

julia> @code_warntype h(x,a,b)
Variables
#self#::Core.Compiler.Const(h, false)
x::Array{Float64,1}
a::Int64
b::Int64
#18::var"#18#19"{Int64,Int64}

Body::Float64
1 ─ %1 = Main.:(var"#18#19")::Core.Compiler.Const(var"#18#19", false)
│   %2 = Core.typeof(a)::Core.Compiler.Const(Int64, false)
│   %3 = Core.typeof(b)::Core.Compiler.Const(Int64, false)
│   %4 = Core.apply_type(%1, %2, %3)::Core.Compiler.Const(var"#18#19"{Int64,Int64}, false)
│        (#18 = %new(%4, a, b))
│   %6 = #18::var"#18#19"{Int64,Int64}
│   %7 = Main.u(%6, x)::Float64
└──      return %7


In this simple example you won't see any measurable performance penalty associated to this type-instability, but certainly that can arise in other examples.