Blog

Articles about computational science and data science, neuroscience, and open source solutions. Personal stories are filed under Weekend Stories. Browse all topics here. All posts are CC BY-NC-SA licensed unless otherwise stated. Feel free to share, remix, and adapt the content as long as you give appropriate credit and distribute your contributions under the same license.

tags ·  RSS ·  Mastodon ·  simple view · page 6/15

 

Bridging ideas on the go: WikiLinks come to DEVONthink To Go

posted:
The WikiLinks feature has finally arrived on DEVONthink to go, DEVONthink’s mobile app, which unleashes new possibilities to work with your Personal Knowledge Management (PKM) system on the go.

New publication on Tauopathy

posted:
A new study on Tauopathy in which our lab was involved has just been published.

Understanding L1 and L2 regularization in machine learning

posted:
Regularization techniques play a vital role in preventing overfitting and enhancing the generalization capability of machine learning models. Among these techniques, L1 and L2 regularization are widely employed for their effectiveness in controlling model complexity. In this blog post, we explore the concepts of L1 and L2 regularization and provide a practical demonstration in Python.

Understanding gradient descent in machine learning

posted:
Gradient descent is a fundamental optimization algorithm widely used in machine learning for finding the optimal parameters of a model. It is a powerful technique that enables models to learn from data by iteratively adjusting their parameters to minimize a cost or loss function. In this blog post, we explore the mathematical background of this method and showcase its implementation in Python.

Loading and saving files in Google Colab

posted:
Enable I/O support in your notebooks running in Google Colab with just a few additional commands.

Mutual information and its relationship to information entropy

posted:
Mutual information is an essential measure in information theory that quantifies the statistical dependence between two random variables. Given its broad applicability, it has become an invaluable tool in diverse fields like machine learning, neuroscience, signal processing, and more. This post explores the mathematical foundations of mutual information and its relationship to information entropy. We will also demonstrate its implementation in some Python examples.

Information entropy

posted:
A fundamental concept that plays a pivotal role in quantifying the uncertainty or randomness of a set of data is the information entropy. Information entropy provides a measure of the average amount of information or surprise contained in a random variable. In this blog post, we explore its mathematical foundations and demonstrate its implementation in some Python examples.

Understanding entropy

posted:
In physics, entropy is a fundamental concept that plays a crucial role in understanding the behavior of physical systems. It provides a measure of the disorder or randomness within a system, and its study has far-reaching applications across various branches of physics. This blog post aims to provide a brief overview of entropy in order to gain a better understanding of it.

Zen and natural sciences

posted: updated:
In this post, I broaden the scope and explore the intersections of Zen and natural sciences more generally.

The Zen of Python

posted:
The connection between Zen and programming is not a subjective one at all. For instance, Python has built it directly into its core programming, known as The Zen of Python.

updated: