Hey Boston Lisp
Geoffrey S. Knauth
geoff at knauth.org
Sat Apr 18 01:33:06 UTC 2020
I first encountered Julia at StrangeLoop a few years ago.
I haven't spent time with it but I intend to look into Julia more.
Do you think there is a possibility that numerical code that would have
been written in Fortran in the past might now be coded or recoded in Julia?
Do Julia macros affect performance in a noticeable way?
Geoff
On Wed, Apr 15, 2020, at 14:05, Michael Bukatin wrote:
>
>
> I have been looking at Julia and its ecosystem in the last few months, and
> it is a very interesting experience. The language has full-strength Lisp
> macros, and full-strength multiple dispatch (so it is a full Lisp), while
> the user-facing syntax is not Lisp-like:
>
> https://docs.julialang.org/en/v1/manual/metaprogramming/
>
> So it's both a Lisp and a non-Lisp.
>
> Generally speaking, people who creat Julia are consistently trying to "eat
> one's cake and to have it too", along multiple dimensions. Another axis is
> that the language is more flexible than Python, but is as fast as C. This
> is achieved via a very tasteful language design (the compiler is a normal
> competent LLVM compiler without miracles, it does not play any special
> role in this combination of expressiveness and speed).
>
> I have also found Julia open-source software on github unusually readable
> and easy to understand (it also tends to be very compact).
>
> The reason I was looking at Julia was that I was having an unusually
> flexible class of machine learning problems (a class of neural machines
> which is based on processing complicated structured data streams, and on
> using "flexible tensors" with tree-shaped indices; so one can do much more
> with these neural machines than with traditional neural nets).
>
> Even the most flexible Python frameworks, such as PyTorch, are too rigid
> for this class of problems, because they are oriented towards fixed
> multidimensional arrays ("tensors").
>
> In this sense, Julia ecosystem seems to have a perfect fit, the Julia Flux
> machine learning framework, which is specifically oriented towards maximal
> flexibility and away from "tensors", while still being focused on high
> performance:
>
> https://github.com/FluxML/Flux.jl
>
> So far I was mostly reading other people's code, and doing small-scale
> explorations of my own (and creating publicly available notes in the
> process): https://github.com/anhinga/2020-julia-drafts
>
> I think that what I am trying to do with Julia Flux should be doable
> single-handedly (the tools seem to be that good), but I also hope to find
> collaborators (a small team would be able to move really fast with this).
>
> - Mishka
>
> On Tue, 14 Apr 2020, Jonathan Godbout wrote:
>
> > Hey Everyone,
> > I hope you're doing well and staying safe.
> > Sorry for the long wait between messages.
> > As Didier just said the ELS will be online, yay!
> >
> > How's everyone doing?
> > Fare, how's the startup, I miss the details.
> >
> > Jon
> >
> >
> >
>
>
--
Geoffrey S. Knauth | http://knauth.org/gsk
More information about the boston-lisp
mailing list