Unified Image and Video Saliency Modeling (ECCV 2020)
Support
Quality
Security
License
Reuse
Predicts Daily NBA Games Using a Logistic Regression Model
Support
Quality
Security
License
Reuse
.Net Building Blocks
Support
Quality
Security
License
Reuse
Repo that relates to the Medium blog 'Keeping your ML model in shape with Kafka, Airflow' and MLFlow'
Support
Quality
Security
License
Reuse
Traditional Machine Learning Models for Large-Scale Datasets in PyTorch.
Support
Quality
Security
License
Reuse
Kafka-ML: connecting the data stream with ML/AI frameworks (now TensorFlow and PyTorch!)
Support
Quality
Security
License
Reuse
The Open Source Computer Aided Dispatch (CAD), Personnel, Shift Management, Automatic Vehicle Location (AVL) and Emergency Management Platform that powers Resgrid.com
Support
Quality
Security
License
Reuse
midi processor library for PerformanceRNN & MusicTransformer published by "Google Magenta"
Support
Quality
Security
License
Reuse
An open source platform for the Industrial Internet of Things, it acts as an edge gateway between sensor devices and cloud storage systems.
Support
Quality
Security
License
Reuse
MLOS is a Data Science powered infrastructure and methodology to democratize and automate Performance Engineering. MLOS enables continuous, instance-based, robust, and trackable systems optimization.
Support
Quality
Security
License
Reuse
d
disentangled_graph_collaborative_filteringby xiangwang1223
Python 88 Version:Current License: No License (No License)
Disentagnled Graph Collaborative Filtering, SIGIR2020
Support
Quality
Security
License
Reuse
Custom error formatting and exception handling in Rest Controllers with Spring Boot
Support
Quality
Security
License
Reuse
:scroll: Visual Studio extension to generate OpenAPI (Swagger) web service reference.
Support
Quality
Security
License
Reuse
Full stack, fully-featured social media web application using React, Firebase, Redux, Express, and Material-UI. Implemented backend REST API server with Node.js and Express, user login and authentication, image uploads, notifications, cloud functions, deploying to Firebase
Support
Quality
Security
License
Reuse
Package normalization ruleset for Repology
Support
Quality
Security
License
Reuse
General Modeling Network Specification
Support
Quality
Security
License
Reuse
P
Python 82 Version:Current License: No License (No License)
Support
Quality
Security
License
Reuse
A
Alzhimers-Disease-Prediction-Using-Deep-learningby himanshub1007
Python 82 Version:Current License: No License (No License)
# AD-Prediction Convolutional Neural Networks for Alzheimer's Disease Prediction Using Brain MRI Image ## Abstract Alzheimers disease (AD) is characterized by severe memory loss and cognitive impairment. It associates with significant brain structure changes, which can be measured by magnetic resonance imaging (MRI) scan. The observable preclinical structure changes provides an opportunity for AD early detection using image classification tools, like convolutional neural network (CNN). However, currently most AD related studies were limited by sample size. Finding an efficient way to train image classifier on limited data is critical. In our project, we explored different transfer-learning methods based on CNN for AD prediction brain structure MRI image. We find that both pretrained 2D AlexNet with 2D-representation method and simple neural network with pretrained 3D autoencoder improved the prediction performance comparing to a deep CNN trained from scratch. The pretrained 2D AlexNet performed even better (**86%**) than the 3D CNN with autoencoder (**77%**). ## Method #### 1. Data In this project, we used public brain MRI data from **Alzheimers Disease Neuroimaging Initiative (ADNI)** Study. ADNI is an ongoing, multicenter cohort study, started from 2004. It focuses on understanding the diagnostic and predictive value of Alzheimers disease specific biomarkers. The ADNI study has three phases: ADNI1, ADNI-GO, and ADNI2. Both ADNI1 and ADNI2 recruited new AD patients and normal control as research participants. Our data included a total of 686 structure MRI scans from both ADNI1 and ADNI2 phases, with 310 AD cases and 376 normal controls. We randomly derived the total sample into training dataset (n = 519), validation dataset (n = 100), and testing dataset (n = 67). #### 2. Image preprocessing Image preprocessing were conducted using Statistical Parametric Mapping (SPM) software, version 12. The original MRI scans were first skull-stripped and segmented using segmentation algorithm based on 6-tissue probability mapping and then normalized to the International Consortium for Brain Mapping template of European brains using affine registration. Other configuration includes: bias, noise, and global intensity normalization. The standard preprocessing process output 3D image files with an uniform size of 121x145x121. Skull-stripping and normalization ensured the comparability between images by transforming the original brain image into a standard image space, so that same brain substructures can be aligned at same image coordinates for different participants. Diluted or enhanced intensity was used to compensate the structure changes. the In our project, we used both whole brain (including both grey matter and white matter) and grey matter only. #### 3. AlexNet and Transfer Learning Convolutional Neural Networks (CNN) are very similar to ordinary Neural Networks. A CNN consists of an input and an output layer, as well as multiple hidden layers. The hidden layers are either convolutional, pooling or fully connected. ConvNet architectures make the explicit assumption that the inputs are images, which allows us to encode certain properties into the architecture. These then make the forward function more efficient to implement and vastly reduce the amount of parameters in the network. #### 3.1. AlexNet The net contains eight layers with weights; the first five are convolutional and the remaining three are fully connected. The overall architecture is shown in Figure 1. The output of the last fully-connected layer is fed to a 1000-way softmax which produces a distribution over the 1000 class labels. AlexNet maximizes the multinomial logistic regression objective, which is equivalent to maximizing the average across training cases of the log-probability of the correct label under the prediction distribution. The kernels of the second, fourth, and fifth convolutional layers are connected only to those kernel maps in the previous layer which reside on the same GPU (as shown in Figure1). The kernels of the third convolutional layer are connected to all kernel maps in the second layer. The neurons in the fully connected layers are connected to all neurons in the previous layer. Response-normalization layers follow the first and second convolutional layers. Max-pooling layers follow both response-normalization layers as well as the fifth convolutional layer. The ReLU non-linearity is applied to the output of every convolutional and fully-connected layer. ![](images/f1.png) The first convolutional layer filters the 224x224x3 input image with 96 kernels of size 11x11x3 with a stride of 4 pixels (this is the distance between the receptive field centers of neighboring neurons in a kernel map). The second convolutional layer takes as input the (response-normalized and pooled) output of the first convolutional layer and filters it with 256 kernels of size 5x5x48. The third, fourth, and fifth convolutional layers are connected to one another without any intervening pooling or normalization layers. The third convolutional layer has 384 kernels of size 3x3x256 connected to the (normalized, pooled) outputs of the second convolutional layer. The fourth convolutional layer has 384 kernels of size 3x3x192 , and the fifth convolutional layer has 256 kernels of size 3x3x192. The fully-connected layers have 4096 neurons each. #### 3.2. Transfer Learning Training an entire Convolutional Network from scratch (with random initialization) is impractical[14] because it is relatively rare to have a dataset of sufficient size. An alternative is to pretrain a Conv-Net on a very large dataset (e.g. ImageNet), and then use the ConvNet either as an initialization or a fixed feature extractor for the task of interest. Typically, there are three major transfer learning scenarios: **ConvNet as fixed feature extractor:** We can take a ConvNet pretrained on ImageNet, and remove the last fully-connected layer, then treat the rest structure as a fixed feature extractor for the target dataset. In AlexNet, this would be a 4096-D vector. Usually, we call these features as CNN codes. Once we get these features, we can train a linear classifier (e.g. linear SVM or Softmax classifier) for our target dataset. **Fine-tuning the ConvNet:** Another idea is not only replace the last fully-connected layer in the classifier, but to also fine-tune the parameters of the pretrained network. Due to overfitting concerns, we can only fine-tune some higher-level part of the network. This suggestion is motivated by the observation that earlier features in a ConvNet contains more generic features (e.g. edge detectors or color blob detectors) that can be useful for many kind of tasks. But the later layer of the network becomes progressively more specific to the details of the classes contained in the original dataset. **Pretrained models:** The released pretrained model is usually the final ConvNet checkpoint. So it is common to see people use the network for fine-tuning. #### 4. 3D Autoencoder and Convolutional Neural Network We take a two-stage approach where we first train a 3D sparse autoencoder to learn filters for convolution operations, and then build a convolutional neural network whose first layer uses the filters learned with the autoencoder. ![](images/f2.png) #### 4.1. Sparse Autoencoder An autoencoder is a 3-layer neural network that is used to extract features from an input such as an image. Sparse representations can provide a simple interpretation of the input data in terms of a small number of \parts by extracting the structure hidden in the data. The autoencoder has an input layer, a hidden layer and an output layer, and the input and output layers have same number of units, while the hidden layer contains more units for a sparse and overcomplete representation. The encoder function maps input x to representation h, and the decoder function maps the representation h to the output x. In our problem, we extract 3D patches from scans as the input to the network. The decoder function aims to reconstruct the input form the hidden representation h. #### 4.2. 3D Convolutional Neural Network Training the 3D convolutional neural network(CNN) is the second stage. The CNN we use in this project has one convolutional layer, one pooling layer, two linear layers, and finally a log softmax layer. After training the sparse autoencoder, we take the weights and biases of the encoder from trained model, and use them a 3D filter of a 3D convolutional layer of the 1-layer convolutional neural network. Figure 2 shows the architecture of the network. #### 5. Tools In this project, we used Nibabel for MRI image processing and PyTorch Neural Networks implementation.
Support
Quality
Security
License
Reuse
MAterials Simulation Toolkit for Machine Learning (MAST-ML)
Support
Quality
Security
License
Reuse
d
deploy-ml-fastapi-redis-dockerby shanesoh
Python 79 Version:Current License: No License (No License)
Deploy and scale machine learning models with FastAPI, Redis and Docker
Support
Quality
Security
License
Reuse
modify self-attention model for EEG signal as input and image embedding layer as output
Support
Quality
Security
License
Reuse
Official version of 'Weakly Supervised 3D object detection from Lidar Point Cloud'(ECCV2020)
Support
Quality
Security
License
Reuse
User's custom models boilerplate
Support
Quality
Security
License
Reuse
Creates CycloneDX Software Bill of Materials (SBOM) from Node.js projects
Support
Quality
Security
License
Reuse
S
SysML-v2-Releaseby Systems-Modeling
JavaScript 78 Version:Current License: Proprietary (Proprietary)
The latest monthly incremental release of SysML v2. Start here.
Support
Quality
Security
License
Reuse
Python Machine Learning Algorithms
Support
Quality
Security
License
Reuse
Bayesian time series forecasting and decision analysis
Support
Quality
Security
License
Reuse
🧬 An R package for visualizing molecular data in 3D
Support
Quality
Security
License
Reuse
Code for the paper "A New Approach to 3D ICP Covariance Estimation for Mobile Robotics"
Support
Quality
Security
License
Reuse
Fully featured .NET library in C# to build MTConnect Agent, Adapter, and Client Applications. Pre-built Agents with Windows Installers. Support for Windows and Linux. Supports MTConnect Versions up to 2.1. Supports .NET Framework 4.6.1 up to .NET 7
Support
Quality
Security
License
Reuse
Graph-based Knowledge Tracing: Modeling Student Proficiency Using Graph Neural Network
Support
Quality
Security
License
Reuse
Support
Quality
Security
License
Reuse
Python port of PeerJS client
Support
Quality
Security
License
Reuse
ML4Chem: Machine Learning for Chemistry and Materials
Support
Quality
Security
License
Reuse
a
aws-appsync-iot-core-realtime-exampleby aws-samples
JavaScript 66 Version:Current License: Permissive (MIT-0)
This sample application demonstrates an iPhone receiving real-time updates from an IoT sensor. The solution is built with AWS AppSync and AWS IoT Core technologies.
Support
Quality
Security
License
Reuse
Simulation and reinforcement learning framework for production planning and control of complex job shop manufacturing systems
Support
Quality
Security
License
Reuse
N
Node-js-functionalitiesby mostlypanda
JavaScript 66 Version:Current License: No License (No License)
This repository contains very useful restful API's and functionalities in node-js containing many important tutorial code for mastering node-js, all tutorials have been published on medium.com, tutorials link is given below
Support
Quality
Security
License
Reuse
LaunchDarkly Server-side SDK for Node
Support
Quality
Security
License
Reuse
Modular tool/armor mod (formerly part of Silent's Gems)
Support
Quality
Security
License
Reuse
Generate G-code to engrave raster images
Support
Quality
Security
License
Reuse
Predict materials properties using only the composition information!
Support
Quality
Security
License
Reuse
I
Image-Classification-by-Keras-and-Tensorflowby rohanpillai20
Python 64 Version:Current License: Permissive (Apache-2.0)
Image Classification using Keras as well as Tensorflow.
Support
Quality
Security
License
Reuse
Official implementation of "Decoder Modulation for Indoor Depth Completion" https://arxiv.org/abs/2005.08607
Support
Quality
Security
License
Reuse
BLE to MQTT Gateway for Smarthome and IoT Devices
Support
Quality
Security
License
Reuse
Crystal Graph Neural Networks
Support
Quality
Security
License
Reuse
Toolkit for Auditing and Mitigating Bias and Fairness of Machine Learning Systems 🔎🤖🧰
Support
Quality
Security
License
Reuse
Soft Threshold Weight Reparameterization for Learnable Sparsity
Support
Quality
Security
License
Reuse
Cyber Attack Detection thanks to Machine Learning Algorithms
Support
Quality
Security
License
Reuse
Laboro BERT Japanese: Japanese BERT Pre-Trained With Web-Corpus
Support
Quality
Security
License
Reuse
M
MPU-6050-9250-I2C-CompFilterby MarkSherstan
Python 61 Version:Current License: Strong Copyleft (GPL-3.0)
MPU6050/9250 I2C and SPI interface. Sensor fusion using a complementary filter yields sensor Euler angles and is implemented in five different languages.
Support
Quality
Security
License
Reuse
u
unisalby rdroste
Unified Image and Video Saliency Modeling (ECCV 2020)
Python 105Updated: 2 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
N
NBA-Predictby JakeKandell
Predicts Daily NBA Games Using a Logistic Regression Model
Python 104Updated: 2 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
n
Support
Quality
Security
License
Reuse
i
incremental_trainingby MBKraus
Repo that relates to the Medium blog 'Keeping your ML model in shape with Kafka, Airflow' and MLFlow'
Python 96Updated: 3 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
p
pycaveby borchero
Traditional Machine Learning Models for Large-Scale Datasets in PyTorch.
Python 93Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
k
kafka-mlby ertis-research
Kafka-ML: connecting the data stream with ML/AI frameworks (now TensorFlow and PyTorch!)
Python 93Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
C
Coreby Resgrid
The Open Source Computer Aided Dispatch (CAD), Personnel, Shift Management, Automatic Vehicle Location (AVL) and Emergency Management Platform that powers Resgrid.com
C# 92Updated: 2 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
m
midi-neural-processorby jason9693
midi processor library for PerformanceRNN & MusicTransformer published by "Google Magenta"
Python 91Updated: 1 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
f
fledgeby fledge-iot
An open source platform for the Industrial Internet of Things, it acts as an edge gateway between sensor devices and cloud storage systems.
Python 90Updated: 1 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
M
MLOSby microsoft
MLOS is a Data Science powered infrastructure and methodology to democratize and automate Performance Engineering. MLOS enables continuous, instance-based, robust, and trackable systems optimization.
Python 89Updated: 1 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
d
disentangled_graph_collaborative_filteringby xiangwang1223
Disentagnled Graph Collaborative Filtering, SIGIR2020
Python 88Updated: 3 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
s
spring-boot-rest-exceptionsby mechero
Custom error formatting and exception handling in Rest Controllers with Spring Boot
Java 86Updated: 2 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
U
Unchase.OpenAPI.Connectedserviceby unchase
:scroll: Visual Studio extension to generate OpenAPI (Swagger) web service reference.
C# 86Updated: 2 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
S
Social-Media-Appby dch133
Full stack, fully-featured social media web application using React, Firebase, Redux, Express, and Material-UI. Implemented backend REST API server with Node.js and Express, user login and authentication, image uploads, notifications, cloud functions, deploying to Firebase
JavaScript 85Updated: 1 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
r
repology-rulesby repology
Package normalization ruleset for Repology
Python 84Updated: 1 y ago License: Strong Copyleft (GPL-3.0)
Support
Quality
Security
License
Reuse
G
GMNSby zephyr-data-specs
General Modeling Network Specification
Jupyter Notebook 84Updated: 1 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
P
Predicting-Myers-Briggs-Type-Indicator-with-Recurrent-Neural-Networksby ianscottknight
Python 82Updated: 2 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
A
Alzhimers-Disease-Prediction-Using-Deep-learningby himanshub1007
# AD-Prediction Convolutional Neural Networks for Alzheimer's Disease Prediction Using Brain MRI Image ## Abstract Alzheimers disease (AD) is characterized by severe memory loss and cognitive impairment. It associates with significant brain structure changes, which can be measured by magnetic resonance imaging (MRI) scan. The observable preclinical structure changes provides an opportunity for AD early detection using image classification tools, like convolutional neural network (CNN). However, currently most AD related studies were limited by sample size. Finding an efficient way to train image classifier on limited data is critical. In our project, we explored different transfer-learning methods based on CNN for AD prediction brain structure MRI image. We find that both pretrained 2D AlexNet with 2D-representation method and simple neural network with pretrained 3D autoencoder improved the prediction performance comparing to a deep CNN trained from scratch. The pretrained 2D AlexNet performed even better (**86%**) than the 3D CNN with autoencoder (**77%**). ## Method #### 1. Data In this project, we used public brain MRI data from **Alzheimers Disease Neuroimaging Initiative (ADNI)** Study. ADNI is an ongoing, multicenter cohort study, started from 2004. It focuses on understanding the diagnostic and predictive value of Alzheimers disease specific biomarkers. The ADNI study has three phases: ADNI1, ADNI-GO, and ADNI2. Both ADNI1 and ADNI2 recruited new AD patients and normal control as research participants. Our data included a total of 686 structure MRI scans from both ADNI1 and ADNI2 phases, with 310 AD cases and 376 normal controls. We randomly derived the total sample into training dataset (n = 519), validation dataset (n = 100), and testing dataset (n = 67). #### 2. Image preprocessing Image preprocessing were conducted using Statistical Parametric Mapping (SPM) software, version 12. The original MRI scans were first skull-stripped and segmented using segmentation algorithm based on 6-tissue probability mapping and then normalized to the International Consortium for Brain Mapping template of European brains using affine registration. Other configuration includes: bias, noise, and global intensity normalization. The standard preprocessing process output 3D image files with an uniform size of 121x145x121. Skull-stripping and normalization ensured the comparability between images by transforming the original brain image into a standard image space, so that same brain substructures can be aligned at same image coordinates for different participants. Diluted or enhanced intensity was used to compensate the structure changes. the In our project, we used both whole brain (including both grey matter and white matter) and grey matter only. #### 3. AlexNet and Transfer Learning Convolutional Neural Networks (CNN) are very similar to ordinary Neural Networks. A CNN consists of an input and an output layer, as well as multiple hidden layers. The hidden layers are either convolutional, pooling or fully connected. ConvNet architectures make the explicit assumption that the inputs are images, which allows us to encode certain properties into the architecture. These then make the forward function more efficient to implement and vastly reduce the amount of parameters in the network. #### 3.1. AlexNet The net contains eight layers with weights; the first five are convolutional and the remaining three are fully connected. The overall architecture is shown in Figure 1. The output of the last fully-connected layer is fed to a 1000-way softmax which produces a distribution over the 1000 class labels. AlexNet maximizes the multinomial logistic regression objective, which is equivalent to maximizing the average across training cases of the log-probability of the correct label under the prediction distribution. The kernels of the second, fourth, and fifth convolutional layers are connected only to those kernel maps in the previous layer which reside on the same GPU (as shown in Figure1). The kernels of the third convolutional layer are connected to all kernel maps in the second layer. The neurons in the fully connected layers are connected to all neurons in the previous layer. Response-normalization layers follow the first and second convolutional layers. Max-pooling layers follow both response-normalization layers as well as the fifth convolutional layer. The ReLU non-linearity is applied to the output of every convolutional and fully-connected layer. ![](images/f1.png) The first convolutional layer filters the 224x224x3 input image with 96 kernels of size 11x11x3 with a stride of 4 pixels (this is the distance between the receptive field centers of neighboring neurons in a kernel map). The second convolutional layer takes as input the (response-normalized and pooled) output of the first convolutional layer and filters it with 256 kernels of size 5x5x48. The third, fourth, and fifth convolutional layers are connected to one another without any intervening pooling or normalization layers. The third convolutional layer has 384 kernels of size 3x3x256 connected to the (normalized, pooled) outputs of the second convolutional layer. The fourth convolutional layer has 384 kernels of size 3x3x192 , and the fifth convolutional layer has 256 kernels of size 3x3x192. The fully-connected layers have 4096 neurons each. #### 3.2. Transfer Learning Training an entire Convolutional Network from scratch (with random initialization) is impractical[14] because it is relatively rare to have a dataset of sufficient size. An alternative is to pretrain a Conv-Net on a very large dataset (e.g. ImageNet), and then use the ConvNet either as an initialization or a fixed feature extractor for the task of interest. Typically, there are three major transfer learning scenarios: **ConvNet as fixed feature extractor:** We can take a ConvNet pretrained on ImageNet, and remove the last fully-connected layer, then treat the rest structure as a fixed feature extractor for the target dataset. In AlexNet, this would be a 4096-D vector. Usually, we call these features as CNN codes. Once we get these features, we can train a linear classifier (e.g. linear SVM or Softmax classifier) for our target dataset. **Fine-tuning the ConvNet:** Another idea is not only replace the last fully-connected layer in the classifier, but to also fine-tune the parameters of the pretrained network. Due to overfitting concerns, we can only fine-tune some higher-level part of the network. This suggestion is motivated by the observation that earlier features in a ConvNet contains more generic features (e.g. edge detectors or color blob detectors) that can be useful for many kind of tasks. But the later layer of the network becomes progressively more specific to the details of the classes contained in the original dataset. **Pretrained models:** The released pretrained model is usually the final ConvNet checkpoint. So it is common to see people use the network for fine-tuning. #### 4. 3D Autoencoder and Convolutional Neural Network We take a two-stage approach where we first train a 3D sparse autoencoder to learn filters for convolution operations, and then build a convolutional neural network whose first layer uses the filters learned with the autoencoder. ![](images/f2.png) #### 4.1. Sparse Autoencoder An autoencoder is a 3-layer neural network that is used to extract features from an input such as an image. Sparse representations can provide a simple interpretation of the input data in terms of a small number of \parts by extracting the structure hidden in the data. The autoencoder has an input layer, a hidden layer and an output layer, and the input and output layers have same number of units, while the hidden layer contains more units for a sparse and overcomplete representation. The encoder function maps input x to representation h, and the decoder function maps the representation h to the output x. In our problem, we extract 3D patches from scans as the input to the network. The decoder function aims to reconstruct the input form the hidden representation h. #### 4.2. 3D Convolutional Neural Network Training the 3D convolutional neural network(CNN) is the second stage. The CNN we use in this project has one convolutional layer, one pooling layer, two linear layers, and finally a log softmax layer. After training the sparse autoencoder, we take the weights and biases of the encoder from trained model, and use them a 3D filter of a 3D convolutional layer of the 1-layer convolutional neural network. Figure 2 shows the architecture of the network. #### 5. Tools In this project, we used Nibabel for MRI image processing and PyTorch Neural Networks implementation.
Python 82Updated: 2 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
M
MAST-MLby uw-cmg
MAterials Simulation Toolkit for Machine Learning (MAST-ML)
Python 80Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
d
deploy-ml-fastapi-redis-dockerby shanesoh
Deploy and scale machine learning models with FastAPI, Redis and Docker
Python 79Updated: 3 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
T
Transformer-for-EEGby redevaaa
modify self-attention model for EEG signal as input and image embedding layer as output
Python 79Updated: 2 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
W
WS3Dby hlesmqh
Official version of 'Weakly Supervised 3D object detection from Lidar Point Cloud'(ECCV2020)
Python 78Updated: 3 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
d
datarobot-user-modelsby datarobot
User's custom models boilerplate
Java 78Updated: 1 y ago License: Proprietary (Proprietary)
Support
Quality
Security
License
Reuse
c
cyclonedx-node-moduleby CycloneDX
Creates CycloneDX Software Bill of Materials (SBOM) from Node.js projects
JavaScript 78Updated: 3 y ago License: Proprietary (Proprietary)
Support
Quality
Security
License
Reuse
S
SysML-v2-Releaseby Systems-Modeling
The latest monthly incremental release of SysML v2. Start here.
JavaScript 78Updated: 3 y ago License: Proprietary (Proprietary)
Support
Quality
Security
License
Reuse
P
Python-Machine-Learningby GeorgeSeif
Python Machine Learning Algorithms
Python 74Updated: 4 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
p
pybatsby lavinei
Bayesian time series forecasting and decision analysis
Jupyter Notebook 74Updated: 2 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
r
r3dmolby swsoyee
🧬 An R package for visualizing molecular data in 3D
JavaScript 73Updated: 2 y ago License: Proprietary (Proprietary)
Support
Quality
Security
License
Reuse
3
3d-icp-covby CAOR-MINES-ParisTech
Code for the paper "A New Approach to 3D ICP Covariance Estimation for Mobile Robotics"
Python 72Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
M
MTConnect.NETby TrakHound
Fully featured .NET library in C# to build MTConnect Agent, Adapter, and Client Applications. Pre-built Agents with Windows Installers. Support for Windows and Linux. Supports MTConnect Versions up to 2.1. Supports .NET Framework 4.6.1 up to .NET 7
C# 71Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
G
GKTby jhljx
Graph-based Knowledge Tracing: Modeling Student Proficiency Using Graph Neural Network
Python 71Updated: 1 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
w
waveplus-readerby Airthings
Python 68Updated: 3 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
p
peerjs-pythonby ambianic
Python port of PeerJS client
Python 68Updated: 2 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
m
ml4chemby muammar
ML4Chem: Machine Learning for Chemistry and Materials
Python 67Updated: 4 y ago License: Proprietary (Proprietary)
Support
Quality
Security
License
Reuse
a
aws-appsync-iot-core-realtime-exampleby aws-samples
This sample application demonstrates an iPhone receiving real-time updates from an IoT sensor. The solution is built with AWS AppSync and AWS IoT Core technologies.
JavaScript 66Updated: 2 y ago License: Permissive (MIT-0)
Support
Quality
Security
License
Reuse
S
SimRLFabby AndreasKuhnle
Simulation and reinforcement learning framework for production planning and control of complex job shop manufacturing systems
Python 66Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
N
Node-js-functionalitiesby mostlypanda
This repository contains very useful restful API's and functionalities in node-js containing many important tutorial code for mastering node-js, all tutorials have been published on medium.com, tutorials link is given below
JavaScript 66Updated: 3 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
n
node-server-sdkby launchdarkly
LaunchDarkly Server-side SDK for Node
JavaScript 65Updated: 2 y ago License: Proprietary (Proprietary)
Support
Quality
Security
License
Reuse
S
Silent-Gearby SilentChaos512
Modular tool/armor mod (formerly part of Silent's Gems)
Java 65Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
r
rastercarveby built1n
Generate G-code to engrave raster images
Python 65Updated: 2 y ago License: Strong Copyleft (GPL-2.0)
Support
Quality
Security
License
Reuse
C
CrabNetby anthony-wang
Predict materials properties using only the composition information!
Python 65Updated: 2 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
I
Image-Classification-by-Keras-and-Tensorflowby rohanpillai20
Image Classification using Keras as well as Tensorflow.
Python 64Updated: 3 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
s
saic_depth_completionby saic-vul
Official implementation of "Decoder Modulation for Indoor Depth Completion" https://arxiv.org/abs/2005.08607
Python 64Updated: 4 y ago License: Weak Copyleft (MPL-2.0)
Support
Quality
Security
License
Reuse
C
Cybeleby Hypfer
BLE to MQTT Gateway for Smarthome and IoT Devices
JavaScript 64Updated: 4 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
c
cgnnby Tony-Y
Crystal Graph Neural Networks
Python 63Updated: 3 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
r
responsiblyby ResponsiblyAI
Toolkit for Auditing and Mitigating Bias and Fairness of Machine Learning Systems 🔎🤖🧰
Python 63Updated: 3 y ago License: Permissive (MIT)
Support
Quality
Security
License
Reuse
S
STRby RAIVNLab
Soft Threshold Weight Reparameterization for Learnable Sparsity
Python 63Updated: 3 y ago License: Permissive (Apache-2.0)
Support
Quality
Security
License
Reuse
C
Cyberattack-Detectionby antoinedelplace
Cyber Attack Detection thanks to Machine Learning Algorithms
Python 62Updated: 2 y ago License: No License (No License)
Support
Quality
Security
License
Reuse
L
Laboro-BERT-Japaneseby laboroai
Laboro BERT Japanese: Japanese BERT Pre-Trained With Web-Corpus
Python 62Updated: 3 y ago License: Proprietary (Proprietary)
Support
Quality
Security
License
Reuse
M
MPU-6050-9250-I2C-CompFilterby MarkSherstan
MPU6050/9250 I2C and SPI interface. Sensor fusion using a complementary filter yields sensor Euler angles and is implemented in five different languages.
Python 61Updated: 2 y ago License: Strong Copyleft (GPL-3.0)
Support
Quality
Security
License
Reuse