#Programmer

R0:7194c033807f71949262faf563aef58e-Scientific Visualization: Python + Matplotlib

Scientific Visualization: Python + Matplotlib

The Python scientific visualisation landscape is huge. It is composed of a myriad of tools, ranging from the most versatile and widely used down to the more specialised and confidential. Some of these tools are community based while others are developed by companies. Some are made specifically for the web, others are for the desktop only, some deal with 3D and large data, while others target flawless 2D rendering.

R0:a56672bf43097f7a50286448af70703b-Isaac Gym: High Performance GPU-Based Physics Simulation For Robot Learning

Isaac Gym: High Performance GPU-Based Physics Simulation For Robot Learning

Isaac Gym offers a high performance learning platform to train policies for wide variety of robotics tasks directly on GPU. Both physics simulation and the neural network policy training reside on GPU and communicate by directly passing data from physics buffers to PyTorch tensors without ever going through any CPU bottlenecks. This leads to blazing fast training times for complex robotics tasks on a single GPU with 2-3 orders of magnitude improvements compared to conventional RL training that uses a CPU based simulator and GPU for neural networks.

R0:fd224a04984225e4bdd2ae7a7e595529-Human Learn

Human Learn

Machine learning covers a lot of ground but it is also capable of making bad decision. We’ve also reached a stage of hype that folks forget that many classification problems can be handled by natural intelligence too. This package contains scikit-learn compatible tools that should make it easier to construct and benchmark rule based systems that are designed by humans. You can also use it in combination with ML models.

R0:e3d9ea294a21c145042e5f31369de739-CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithms

CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithms

CARLA (Counterfactual And Recourse LibrAry), a python library for benchmarking counterfactual explanation methods across both different data sets and different machine learning models. In summary, our work provides the following contributions: (i) an extensive benchmark of 11 popular counterfactual explanation methods, (ii) a benchmarking framework for research on future counterfactual explanation methods, and (iii) a standardized set of integrated evaluation measures and data sets for transparent and extensive comparisons of these methods. We have open-sourced CARLA and our experimental results on Github, making them available as competitive baselines. We welcome contributions from other research groups and practitioners.

R0:2271580d5e4f655f084ee4605a50b147-labml.ai Deep Learning Paper Implementations

labml.ai Deep Learning Paper Implementations

This is a collection of simple PyTorch implementations of neural networks and related algorithms. These implementations are documented with explanations, and the website renders these as side-by-side formatted notes. We believe these would help you understand these algorithms better.

R0:3d92323b5375746d21dcb172e8950adc-Explainability in Graph Neural Networks: A Taxonomic Survey

Explainability in Graph Neural Networks: A Taxonomic Survey

We summarize current datasets and metrics for evaluating GNN explainability. Altogether, this work provides a unified methodological treatment of GNN explainability and a standardized testbed for evaluations.

https://editorialia.com/wp-content/uploads/2020/06/openmined-opensource-to-make-privacy-preserving-of-ai-technologies.jpg

OpenMined: open source to make privacy-preserving of AI technologies

With OpenMined, an AI model can be governed by multiple owners and trained securely on an unseen, distributed dataset.The mission of the OpenMined community is to create an accessible ecosystem of tools for private, secure, multi-owner governed AI