This week @NVIDIA open sourced the 3D object generation AI model, GET3D. GET3D is a generative model of high quality 3D textured shapes learned from images.
1/4
Trained using only 2D images, GET3D generates 3D shapes with high-fidelity textures and complex geometric details.
2/4
These 3D objects are created in the same format used by popular graphics software applications, allowing users to immediately import their shapes into 3D renderers and game engines for further editing.
3/4
I have just done something really cool - I've managed to *train* XGBoost in browser completely within an HTML file! This has been possible thanks to the PyScript project that allows running Python inside of HTML, similar to how JavaScript works.
The example below is very simple - the script loads the small Iris dataset from sklearn. With a slider you are able to adjust the number of XGBoost trees, and the script will train different XGBoost models accordingly and print out accuracy.
2/5
PyScript is still in very early stages of development. Getting all the relevant components to work together is still tricky, and there are not many detailed tutorials. Hence, this example is *very* rudimentary. I'll try to make it more powerful and snazzy does the road.
3/5
First, you create a simple model with XGBoost. It doesn't have to be fancy, or even too accurate, it's just for reference purposes. Use that model to calculate the Shapley values for your training set. Here is an example:
NVIDIA GTC starts today! There are tons of exciting topics and webinars covered. This year again the whole conference is online and free, so go and register if you have not done so already.
A very good paper I came across this morning by the @DeepMind researchers. For the past five years Transformers have been one of the most dominant approaches to Deep Learning problems, especially in the #NLP domain.
1/5
However, despite many interesting papers on the topic, and lots of good open code, there has been a noticeable lack of *formal* definition of what transformed are, especially on the level of pseudocode.
2/5
This paper aims to rectify that. It provides pseudocode for almost all major Transformer architectures, including training algorithms.
3/5
The longer you work with ML algorithms, the more you appreciate what an outsize effect your *data* has on the quality of your models. I've seen that shift on Kaggle over the years, where more and more time is spent on some kind of dataset augmentation.
1/5
There is still only so much you can do there, and unless you are "enterprising" and decide to scrape the competition host's website for their dat (yes, this has happened) your legitimate options are rather limited.
2/5
Outside of the Kaggle world, however, things are different. Large computational resources and advanced algorithms still dominate the ML discourse, but those who are paying attention know that neither of them would be worth much without the huge datasets that are bing used.
3/5
It's actually scary how ignorant academics who try to do research on NNs for tabular data are of tabular data. I think that part of the problem is that almost all of the interesting and relevant tabular data problems are in the industry, 1/3
and academics tend to be completely inured from any kind of practical application of ML/DS.
If you are an academic who is interested in doing research on tabular data,
2/3
I would BEG YOU, FOR THE LOVE OF EVERYTHING THAT IS DECENT, PLEASE, PLEASE PLEASE GET OUT OF YOUR IVORY TOWER AND TRY TO LEARN WHAT KINDS OF PROBLEMS ACTUAL DATA SCIENTISTS DEAL WITH IN THEIR PROFESSIONAL LIVES!!!
3/3