Programación y frameworks en IA | 🇬🇧 Programming and frameworks in AI

Auto Added by WPeMatico

Sequence Feature Extraction for Malware Family Analysis via Graph Neural Network

Malicious software (malware) causes much harm to our devices and life. We are eager to understand the malware behavior and the threat it made. Most of the record files of malware are variable length and text-based files with time stamps, such as event log data and dynamic analysis profiles. Using the time stamps, we can […]

Sequence Feature Extraction for Malware Family Analysis via Graph Neural Network Read More »

Scientific Visualization: Python + Matplotlib

The Python scientific visualisation landscape is huge. It is composed of a myriad of tools, ranging from the most versatile and widely used down to the more specialised and confidential. Some of these tools are community based while others are developed by companies. Some are made specifically for the web, others are for the desktop

Scientific Visualization: Python + Matplotlib Read More »

Between words and characters: A Brief History of Open-Vocabulary Modeling and Tokenization in NLP

In this survey, we connect several lines of work from the pre-neural and neural era, by showing how hybrid approaches of words and characters as well as subword-based approaches based on learned segmentation have been proposed and evaluated. We conclude that there is and likely will never be a silver bullet singular solution for all

Between words and characters: A Brief History of Open-Vocabulary Modeling and Tokenization in NLP Read More »

Federated Learning: Issues in Medical Application

In this presentation, the current issues to make federated learning flawlessly useful in the real world will be briefly overviewed. They are related to data/system heterogeneity, client management, traceability, and security. Also, we introduce the modularized federated learning framework, we currently develop, to experiment various techniques and protocols to find solutions for aforementioned issues. The

Federated Learning: Issues in Medical Application Read More »

Auto Quantum Circuits

«AutoQML, self-assembling circuits, hyper-parameterized Quantum ML platform, using cirq, tensorflow and tfq. Trillions of possible qubit registries, gate combinations and moment sequences, ready to be adapted into your ML flow. Here I demonstrate climatechange, jameswebbspacetelescope and microbiology vision applications… [Thus far, a circuit with 16-Qubits and a gate sequence of [ YY ] – [ XX ] – [CNOT]

Auto Quantum Circuits Read More »

Isaac Gym: High Performance GPU-Based Physics Simulation For Robot Learning

Isaac Gym offers a high performance learning platform to train policies for wide variety of robotics tasks directly on GPU. Both physics simulation and the neural network policy training reside on GPU and communicate by directly passing data from physics buffers to PyTorch tensors without ever going through any CPU bottlenecks. This leads to blazing

Isaac Gym: High Performance GPU-Based Physics Simulation For Robot Learning Read More »

Introduction to Datascience: Learn Julia Programming, Math & Datascience from Scratch

I was emboldened to write this book after my video series called Data Science With Julia got some traction. That too after a tweet about Decision Tree was liked by Julia Language itself. So I thought why not give it more?

Introduction to Datascience: Learn Julia Programming, Math & Datascience from Scratch Read More »

Model-based Decision Making with Imagination for Autonomous Parking

Autonomous parking technology is a key concept within autonomous driving research. This paper will propose an imaginative autonomous parking algorithm to solve issues concerned with parking.

Model-based Decision Making with Imagination for Autonomous Parking Read More »

Do Vision Transformers See Like Convolutional Neural Networks?

Convolutional neural networks (CNNs) have so far been the de-facto model for visual data. Recent work has shown that (Vision) Transformer models (ViT) can achieve comparable or even superior performance on image classification tasks. This raises a central question: how are Vision Transformers solving these tasks? Are they acting like convolutional networks, or learning entirely

Do Vision Transformers See Like Convolutional Neural Networks? Read More »

YOLOX: Exceeding YOLO Series in 2021

We switch the YOLO detector to an anchor-free manner and conduct other advanced detection techniques, i.e., a decoupled head and the leading label assignment strategy SimOTA to achieve state-of-the-art results across a large scale range of models: For YOLO-Nano with only 0.91M parameters and 1.08G FLOPs, we get 25.3% AP on COCO, surpassing NanoDet by

YOLOX: Exceeding YOLO Series in 2021 Read More »