All Projects → goru001 → Code With Ai

goru001 / Code With Ai

Licence: mit
Interface for people to use my model which predicts which techniques one should use to solve a competitive programming problem to get an AC

Projects that are alternatives of or similar to Code With Ai

Spark Tpc Ds Performance Test
Use the TPC-DS benchmark to test Spark SQL performance
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Dataminingnotebooks
This is a collection of iPython notebooks from my course on data mining. Data used in the notebooks can be downloaded from the given links in the notebooks.
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Making with ml
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Srn multilabel
Stars: ✭ 132 (-1.49%)
Mutual labels:  jupyter-notebook
Learning Data Mining With Python
Code repo for Learning Data Mining with Python, published by Packt Publishing
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Computational Statistics With Python
Resources for STA 633 class
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Ml Class
Machine learning lessons and teaching projects designed for engineers
Stars: ✭ 1,743 (+1200.75%)
Mutual labels:  jupyter-notebook
Depth Hints
[ICCV 2019] Depth Hints are complementary depth suggestions which improve monocular depth estimation algorithms trained from stereo pairs
Stars: ✭ 134 (+0%)
Mutual labels:  jupyter-notebook
Laplacians.jl
Algorithms inspired by graph Laplacians: linear equation solvers, sparsification, clustering, optimization, etc.
Stars: ✭ 132 (-1.49%)
Mutual labels:  jupyter-notebook
Tplinker Joint Extraction
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Pytorch Tutorials
Pytorch tutorials for Neural Style transfert
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Clockwork Fcn
Clockwork Convnets for Video Semantic Segmenation
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Streaming Demos
Demos of Plotly's Real-time Streaming API
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Yolo Powered robot vision
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Bcs workshop apr 20
Workshop on basic machine learning, computational modeling, psychophysics, basic data analysis and experiment design
Stars: ✭ 134 (+0%)
Mutual labels:  jupyter-notebook
Mathstatscode
Codes for my mathematical statistics course
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Poppy Ergo Jr
🤖 Poppy Ergo Jr is an open-source robotic arm based on modular 3D printed conception and low-cost XL-320 motors.
Stars: ✭ 133 (-0.75%)
Mutual labels:  jupyter-notebook
Machine Learning Concepts
Machine Learning Concepts with Concepts
Stars: ✭ 134 (+0%)
Mutual labels:  jupyter-notebook
Readata
Python数据分析实战项目汇总~
Stars: ✭ 134 (+0%)
Mutual labels:  jupyter-notebook
Fastdoc
Create publication-quality books from Jupyter notebooks
Stars: ✭ 134 (+0%)
Mutual labels:  jupyter-notebook

Code with AI

Code with AI is a tool which predicts which techniques one should use to solve a competitive programming problem to get an AC

Tool: https://code-with-ai.app.render.com/

Usage:

  1. Copy paste the problem statement
  2. Hit Enter :)

Alt Text

  • This tool was very well received by competitive programming community. Check out my blog post on codeforces here

  • This tool was also shared by Jeremy Howard on Twitter. Check out his tweet.

  • Some Analytics: 3300+ unique users have used this tool, till June 2019 since its launch.

Dataset

I scraped the dataset from Codechef and Codeforces, the scripts for which along with the dataset you can find in ulmfit-model/competitive-programming-websites-scripts

The dataset on which model has been trained has 92 classes, which are:

'mobius-function', 'queue', 'greedy', 'suffix-array', 'bitmasks', 'digit dp', 'lca', 'gcd', 'probabilities', 'combinatorics', 'graph matchings', 'easy-medium', 'precomputation', 'sprague-grundy', 'math', 'centroid-decomposition', 'link-cut-tree', 'expression parsing', 'constructive algorithms', 'medium', 'schedules', 'euler tour', 'easy', 'challenge', 'implementation', 'binary search', 'matrices', 'two pointers', 'dfs', 'dp+bitmask', 'sets', 'sqrt-decomposition', 'dijkstra', 'line-sweep', 'data structures', 'tree-dp', 'hard', 'mst', 'recursion', 'games', 'suffix-trees', 'kmp', 'stack', 'brute force', 'medium-hard', 'prefix-sum', 'graphs', '2-sat', 'shortest paths', 'heavy-light', 'heaps', '*special problem', 'trees', 'array', 'sliding-window', 'inclusion-exclusion', 'meet-in-the-middle', 'dfs and similar', 'sortings', 'pigeonhole', 'xor', 'gaussian-elimination', 'lucas theorem', 'divide and conquer', 'flows', 'strings', 'matrix-expo', 'number theory', 'bipartite', 'knapsack', 'sieve', 'ternary search', 'modulo', 'backtracking', 'treap', 'trie', 'dp', 'fenwick', 'observation', 'fibonacci', 'convex-hull', 'chinese remainder theorem', 'string suffix structures', 'geometry', 'lazy propagation', 'factorization', 'dsu', 'fft', 'segment-tree', 'hashing', 'bfs', 'prime', 'mo algorithm'

These classes are a union of tags which codeforces and codechef have! You can check out more details about data preparation in ulmfit-model/code-with-ai-data-preparation.ipynb

Model

You can check out the code for model in notebook ulmfit-model/code-with-ai.ipynb

Basically, the classifier is built starting from Wikitext103 Language Model pretrained weights, fine-tuning the language model on current dataset and then building classifier on top of fine-tuned language model.

Results

The classifier has a F1 Score of ~49.

TODOs

  • Try improving the score further by using bidirectional RNN

  • Try improving the score further by using an approach similar to DeViSE paper, i.e instead of training the model to predict O or 1, train the model to go closer towards the embedding vector representation of labels - the intuition behind this is that labels in competitive programming like graph, dfs, bfs etc aren’t disjoint.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].