(date: 2024-09-01 11:20:27)
date: 2024-09-02, from: ETH Zurich, recently added
Wicki, Juliette
http://hdl.handle.net/20.500.11850/691714
date: 2024-09-01, from: ETH Zurich, recently added
Ho L.; Pham K.; Barthel M.; Harris S.; Bodé S.; De Vrieze J.; Vermeir P.; Six J.; Boeckx P.; Goethals P.
http://hdl.handle.net/20.500.11850/678459
date: 2024-09-01, from: ETH Zurich, recently added
Akindele N.; Taiwo R.; Sarvari H.; Oluleye B.I.; Awodele I.A.; Olaniran T.O.
http://hdl.handle.net/20.500.11850/678468
date: 2024-09-01, from: ETH Zurich, recently added
Rohr-Fukuma M.; Stieglitz L.H.; Bujan B.; Jedrysiak P.; Oertel M.F.; Salzmann L.; Baumann C.R.; Imbach L.L.; Gassert R.; Bichsel O.
http://hdl.handle.net/20.500.11850/680808
date: 2024-09-01, from: ETH Zurich, recently added
Ortmann L.; Böhm F.; Klein-Helmkamp F.; Ulbig A.; Bolognani S.; Dörfler F.
http://hdl.handle.net/20.500.11850/680806
date: 2024-09-01, from: ETH Zurich, recently added
Gong, Congcong; Luan, Xiaojie; Yang, Lixing; Qi, Jianguo; Corman, Francesco
http://hdl.handle.net/20.500.11850/685514
date: 2024-09-01, from: ETH Zurich, recently added
Liu, Yujia; Obukhov, Anton; Wegner, Jan Dirk; Schindler, Konrad
http://hdl.handle.net/20.500.11850/686642
date: 2024-09-01, from: ETH Zurich, recently added
Beutler, Philipp; Larsen, Tove A.; Maurer, Max; Staufer, Philipp; Lienert, Judit
http://hdl.handle.net/20.500.11850/687981
date: 2024-09-01, from: ETH Zurich, recently added
Nickel I.; Potapov E.; Sun B.; Zimpfer D.; Koliopoulou A.; Adachi I.; Anyanwu A.; Falk V.; Atluri P.; Faerber G.; Goldstein D.; Yarboro L.; Slaughter M.S.; Milano C.; Tsukashita M.; D’Alessandro D.; Silvestry S.; Kirov H.; Bommareddi S.; Lanmüller P.
http://hdl.handle.net/20.500.11850/688257
date: 2024-09-01, from: ETH Zurich, recently added
Schöni, Lorin; Carles, Victor; Strohmeier, Martin; Mayer, Peter; Zimmermann, Verena
http://hdl.handle.net/20.500.11850/689291
date: 2024-08-29, from: Standard Ebooks, new releaases
A brother and sister, made destitute by the death of their father, struggle to succeed despite the callous indifference of their corrupt uncle.
https://standardebooks.org/ebooks/charles-dickens/nicholas-nickleby
date: 2024-08-29, from: Standard Ebooks, new releaases
A university student in St. Petersburg is drawn into the activities of a revolutionary group against his will.
https://standardebooks.org/ebooks/joseph-conrad/under-western-eyes
date: 2024-08-29, from: Standard Ebooks, new releaases
A collection of poems by W. B. Yeats.
https://standardebooks.org/ebooks/w-b-yeats/poetry
date: 2024-08-29, from: Standard Ebooks, new releaases
A collection of poems by Frances Ellen Watkins Harper.
https://standardebooks.org/ebooks/frances-ellen-watkins-harper/poetry
date: 2024-08-28, from: Bruno Rodrigues blog
Nix is not only a package manager, but also a build automation tool, and you can use it to build polyglot data science pipelines in a completely reproducible way.
For example, suppose that you need to mix Python, R and maybe some others tools for a project (by the way, some believe this will become the norm in the coming years, use your favourite search engine to look for “polyglot data science” and you’ll see), and suppose that you want to define your project as a nice reproducible pipeline, and not simply a series of scripts. What are the options available to you?
One option would be to use the {targets}
package for R,
which allows you to do lay out your project as pipeline. But as amazing
as {targets}
is, it only works with R. If you also need
Python, you would then need to also use the {reticulate}
package to interface with it. But what do you do if you need some other
command line tools? Well, you could wrap them in an R function using
system()
or system2()
. But what if you need
yet another language, like Julia? There might be a way to call Julia
from R, but as you see, the more diverse tools you need, the more
complex it gets. And it doesn’t really matter if you switch from
{targets}
to another such package that exists for, say,
Python, you would always need to write wrappers or use packages that
allow you to call the other programming languages that you need.
Another possibility is to use good old make
.
make
is a tool from the GNU project that allows you to
define targets, which would be the outputs of a script or call
to some cli tool by writing so-called Makefiles
. For an
example of a Makefile
in research, take a look at
this
one from a
paper
by Grant McDermott. You can
use make
as a to orchestrate several programming languages
or cli tools, but you will need to write code to pass data from one
script to the other. {targets}
deals with that
transparently by serialising all the targets’ outputs using
saveRDS()
but this only works because only R is supported.
But if you’re trying to make R, Python, and whatever else work together,
you will need to deal with this manually and find a common interface to
pass data around.
Despite this, using make
, or some other tool on top of the
required programming languages (and not tied to either one), is likely
the best solution and it turns out that Nix can be used just like that!
But why use Nix and not make
then? Well, using Nix
guarantees that whatever you produce will be completely reproducible.
With make
, you would need to either run it inside a Docker
image or… inside a development environment built with Nix! I did
something similar in
this
blog post where I ran a {targets}
pipeline inside a Nix
environment to make the analysis reproducible.
But if I’m already defining a reproducible development environment using Nix, why not go all the way and build a complete project using Nix? After all, Nix allows you to package software and what is software but 0’s and 1’s? And what is a trained model, a paper or report in the PDF format, predictions exported into a CSV file, etc, if not 0’s and 1’s?
Just like with any other build automation tool, Nix will only rebuild
the project if something changes, and will only rebuild the parts that
need to be rebuilt. So if you change a file somewhere, only whatever
depends on this file will get rebuilt, just like with
{targets}
, or make
.
In the following repository you can find an example of this.
This is a very simple project: two functions are defined in the
python_functions.py
script. These functions are nothing
special, and could be used interactively. One function reads a
.csv
file from the Internet and returns it, the other does
some basic cleaning. Here are these two functions included in the
python_functions.py
file:
from pandas import read_csv
def download_iris(iris_csv_url):
# Read the CSV file
df = read_csv(iris_csv_url)
return df
def process_iris(iris_csv_path):
# Read the CSV file
df = read_csv(iris_csv_path)
# Replace the species numbers with their corresponding names
species_mapping = {0: "setosa", 1: "virginica", 2: "versicolor"}
df['species'] = df['species'].replace(species_mapping)
return df
Then, I want to use {ggplot2}
to plot this data. You will
notice the lack of R script in the repo. I did this on purpose, because
I wanted to show how you could directly write R code inside of a Nix
expression. But in practice, it is better to have Python code in a
Python script, R code in an R script, and then use Nix to orchestrate
the whole thing. But I just wanted to show you that you could, if you
wanted to, have a completely self-contained Nix expression that
encapsulates the business logic as well.
There’s also a .Qmd
file: this is the file that will get
compiled into a PDF document, and is the output of the whole project. It
could be anything else! As I stated above, this is just 0’s and 1’s so
it could very well be some other output, it doesn’t really matter.
Let’s now take a look at the default.nix
that builds the
whole thing. Let’s start by the top-level definitions:
let
pkgs =
import
(fetchTarball "https://github.com/NixOS/nixpkgs/archive/27285241da3bb285155d549a11192e9fdc3a0d04.tar.gz")
{ };
tex = (
pkgs.texlive.combine {
inherit (pkgs.texlive) scheme-small;
}
);
# Because building happens in sandbox that cannot connect to the internet
# we need to download assets beforehand
iris_path = pkgs.fetchurl {
url = "https://raw.githubusercontent.com/b-rodrigues/nixbat/7c319bcdbe15e7f7182e7685b8de176a40d0bde9/iris.csv";
hash = "sha256-2H6THCXKxIt4yxnDDY+AZRmbxqs7FndCp4MqaAR1Cpw=";
};
# Common python dependencies to use in my intermediary inputs
pythonEnv = pkgs.python312.withPackages (ps: with ps; [ pandas ]);
# Common python sources
python_src = pkgs.lib.fileset.toSource {
root = ./.;
fileset = ./python_functions.py;
};
Some variables are defined there:
pkgs
: this is the set of Nix packages to be used. All the
dependencies of the project will get built using the Nix expressions
available in the nixpkgs
Github repository at a specific
commit. This ensures that the output of this expression will always be
exactly the same.
tex
: defines the set of LaTeX packages I need to compile
the PDF.
iris_path
: the Python function I use to load the data takes
a path, or url, to read the iris dataset. Because building a derivation
happens in a sandbox, I need to download assets beforehand. This is what
the fetchurl
function does. I can then refer to the file
path using ${iris_path}
later on.
pythonEnv
: This lists the dependencies I will need to run
my Python functions.
pythonSrc
: Defines the path to the
python_functions.py
file.
Then, I want to call each of my functions separately, and I want them to produce a single output. So for this, I now build a derivation, one per output. I start with the first one:
downloadCsv = pkgs.stdenv.mkDerivation {
name = "download-csv";
buildInputs = [ pythonEnv ];
src = pythonSrc;
buildPhase = ''
python -c "
import pandas as pd
from python_functions import download_iris
iris_raw = download_iris('${iris_path}')
iris_raw.to_csv('iris_raw.csv', index=False)
"
'';
installPhase = ''
mkdir -p $out
cp iris_raw.csv $out/
'';
};
At first sight, there might seem that a lot is going on, but let’s take a closer look:
name = “download-csv”
buildInputs
. This is
what’s required to build the target!
python_functions.py
file
Then, I need to run the code, and this is what happens in the
buildPhase
. This is exactly the code you would write if you
were using a script to glue your functions together. See how I use
irispath < /code > torefertothepathtothefiledefinedabove.Finally, inthe < code > installPhase < /code > Icopythe < code > .csv < /code > fileto < code>out/
,
which essentially copies the file into the Nix store, making it
available for the next derivations.
In the next derivation, I now use the second Python function to clean the data:
cleanCsv = pkgs.stdenv.mkDerivation {
name = "clean-csv";
buildInputs = [ pythonEnv ];
src = pythonSrc;
buildPhase = ''
python -c "
import pandas as pd
from python_functions import process_iris
iris = process_iris('${downloadCsv}/iris_raw.csv')
iris.to_csv('iris.csv', index=False)
"
'';
installPhase = ''
mkdir -p $out
cp iris.csv $out/
'';
};
This is not very different than what I did before. Just notice how I
refer to the output of the first derivation:
${downloadCsv}/iris_raw.csv
.
Now comes the last intermediary derivation, the one that uses R to create a plot:
generatePlot = pkgs.stdenv.mkDerivation {
name = "generate-plot";
buildInputs = with pkgs; [
R
rPackages.ggplot2
rPackages.janitor
];
dontUnpack = true;
buildPhase = ''
Rscript -e "
library(ggplot2)
library(janitor)
iris <- read.csv('${cleanCsv}/iris.csv') |>
clean_names() |>
transform(species = as.character(species))
p <- ggplot(iris,
aes(x = sepal_length, y = sepal_width, color = species)) +
geom_point(size = 3) +
labs(title = 'Sepal Length vs Sepal Width',
x = 'Sepal Length',
y = 'Sepal Width') +
theme_minimal() +
theme(plot.title = element_text(hjust = 0.5))
ggsave('plot.png', plot = p, width = 6, height = 4, dpi = 300)
"
'';
installPhase = ''
mkdir -p $out
cp plot.png $out/
'';
};
As I said above, to make this better, it would need to be a function defined in its own R script, as this way there’s a nice separation of concerns. On one hand, there’s the business logic in Python and R scripts, and on the other there’s the orchestration in Nix. Putting R code in the Nix expression makes this less flexible, but I wanted to show you that this is also a possibility!
Now comes the last part of the Nix expression, the actual thing I want to build, a PDF that uses the generated plot as an input:
in
# Derivation to generate the PDF report from Markdown
pkgs.stdenv.mkDerivation {
name = "generate-report";
buildInputs = [
pkgs.quarto
tex
];
src = pkgs.lib.fileset.toSource {
root = ./.;
# Only include report.Qmd in the source
fileset = ./report.Qmd;
};
buildPhase = ''
cp ${generatePlot}/plot.png .
# Deno needs to add stuff to $HOME/.cache
# so we give it a home to do this
mkdir home
export HOME=$PWD/home
quarto render report.Qmd --to pdf
'';
installPhase = ''
mkdir -p $out
cp report.pdf $out/
'';
}
Notice the dependencies of this derivation: quarto
and
tex
(tex
is the variable I defined right at
the beginning that lists LaTeX packages). I then need to specify
report.Qmd
as the source of this derivation, and copy the
plot generated before in R into the working/build directory. There’s
also a idiosyncrasy where a dependency of Quarto, Deno, needs to have a
directory to save some stuff in it. Nix being Nix, we need to manually
define such a home directory for reproducibility purposes. If it would
be using my home/
directory on my machine, this wouldn’t be
reproducible! We finish the buildPhase
by rendering the
document, and then install it into $out/
. To build
this project, you need to have Nix installed and then type
nix-build
, or alternatively, nix-build -Q
which hides all the output of the build phases (so you don’t see any
warnings or messages thrown by either Python or R).
This will build the PDF, which you can then find in the Nix store.
You’ll notice a file called result
appear next to all your
other files from the project. In a terminal, call readlink
result
and this will show you the path to the generated PDF,
which you can now read!
In conclusion, I think that this is a really useful way to orchestrate
code written in different programming languages, but I would not use
this for monolingual projects. For R, I’ll keep using
{targets}
together with a Nix shell to ensure
reproducibility. Also, to really benefit from this, your code needs,
ideally, to be written as a series of functions, each outputting a
single object. Instead, if you write a script to orchestrate the whole
thing in R or Python, and then put a Nix expression on top of it, I’m
not sure it’s really worth it. Might as well just use a Nix shell then
and execute your scripts in it.
Also, let me state that this is my first attempt at using Nix for such a purpose, and there might be a better/more elegant way of doing it, so if you have any input, don’t hesitate!
Thanks to the amazing Nix community for helping out!
Hope you enjoyed! If you found this blog post useful, you might want to follow me on Mastodon or twitter for blog post updates and buy me an espresso or paypal.me, or buy my ebooks. You can also watch my videos on youtube. So much content for you to consoom!
https://www.brodrigues.co/blog/2024-08-27-nix_for_r_part_12/
date: 2024-08-26, from: Scholarly Kitchen
We’re taking the last week of August off and will be back after the Labor Day holiday.
The post Off Through Labor Day appeared first on The Scholarly Kitchen.
https://scholarlykitchen.sspnet.org/2024/08/26/off-through-labor-day/