Tristan Profile picture
24 Nov, 11 tweets, 6 min read
Curious what Tesla means by upreving their static obstacle neural nets?

Lets see how the Tesla FSD Beta 10.5 3D Voxel nets compare to the nets from two months ago.

The new captures are from the same area as the old ones so we can directly compare the outputs

1/N
This first example is a small pedestrian crosswalk sign in the middle of the road. It's about 1 foot wide so it should show up as 1 pixel in the nets.

Under the old nets it shows up as a large blob with an incorrect depth. Under the new nets it's much better. crosswalk sign in the middle of streetcone under old nets is a large blobcone under new nets is correctly sized to reality
Under the old nets the posts show up a huge blobs and disappears when the car gets close to it. The probabilities seem fairly consistent no matter how far they sign is away even though up close they should be more confident.

fn.lc/s/depthrender/…
Under 10.5 the post has a much more accurate depth up close and the post now shows up along side the car and behind it

The confidence seems to match reality much better as the predictions have a higher probability as the car gets closer

fn.lc/s/depthrender/…
Looking at an intersection near a park we can see that there's a huge difference in clarity.

It seems like Tesla has optimized the training data to better isolate the sidewalks from the ground so they more consistently show up

Old: fn.lc/s/depthrender/… old output10.5 output
The light poles also have much more accurate depth. Before they would have the same issue of having the right width but show up as being 10+ ft deep

New: fn.lc/s/depthrender/…
I don't have an old capture of these cones to compare but they seem plenty distinct and good enough to drive on. The predictions are stable next to and behind the car

Of note is that the cones merge together into a single line further way from the car a city street with lots of cones narrowing the street down ta birdseye view of the cones showing them individually up cla view of the cones in front and behind the car
Since I suspect this training data is automatically generated, the repetitive colors and patterns of the cones may be confusing their offline algorithm into thinking it's one solid object.

I'm not sure it matters for driving given their close spacing

fn.lc/s/depthrender/…
Here's an example of making a right turn into a narrow road with hard curbs on both sides. The car is outputting pretty accurate estimates of the curbs through the tight corner

Seems like my added car model is a tad too far forward compared to reality

fn.lc/s/depthrender/…
It's pretty clear that there's been some big though incremental improvements to these nets.

The recent FSD patch notes mentioning the increased number of clips as well as improvements to the training data generation (autolabeler?) seems to be paying off
It's not clear if this is part of the autolabeler given there's no real labels here but if it's a shared system for managing the clips it may benefit both

I'm also curious if there's been any model architectural changes helping though the patch notes haven't mentioned that

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Tristan

Tristan Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @rice_fry

11 Oct
Tesla has added new voxel 3D birdseye view outputs and it's pretty amazing!

Nice of them to start merging some bits of FSD into the normal firmware in 2021.36 so we can play with the perception side 🙂

Thanks to @greentheonly for the help!
Most of the critical FSD bits are missing in the normal firmware. These outputs aren't normally running but with some tricks we can enable it.

This seems to be the general solution to handling unpredictable scenarios such as the Seattle monorail pillars or overhanging shrubbery.
The nets predict the location of static objects in the space around them via a dense grid of probabilities.

The output is a 384x255x12 dense grid of probabilities. Each cube seems to be ~0.33 meters and currently outputs predictions ~100 meters in front of the vehicle.
Read 15 tweets
12 Apr
We recently got some insight into how Tesla is going to replace radar in the recent firmware updates + some nifty ML model techniques

⬇️ Thread
From the binaries we can see that they've added velocity and acceleration outputs. These predictions in addition to the existing xyz outputs give much of the same information that radar traditionally provides
(distance + velocity + acceleration).
For autosteer on city streets, you need to know the velocity and acceleration of cars in all directions but radar is only pointing forward. If it's accurate enough to make a left turn, radar is probably unnecessary for the most part.
Read 15 tweets
10 Apr
Got a sample of the Tesla Insurance telemetry data. The insurance records are on a per drive basis. Here's the fields:

* Unique Drive ID
* Record Version
* Car Firmware Version
* Driver Profile Name
* Start / End Time
* Drive Duration
* Start / End Odometer

(1/2)
* # of Autopilot Strikeouts
* # of Forward Collision Warnings
* # of Lane Departure Warnings
* # of ABS activations (All & User)
* Time spent within 1s of car in front
* Time spent within 3s of car in front
* Acceleration Variance
* Service Mode
* Delivered

(2/2)
There's lot of basic stuff which insurance companies can get via companion apps/dongles but there's a lot of deep insights into driver behavior which Tesla can get but others cannot.

I bet a lot of insurance companies would love to get their hands on this kind of data
Read 11 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Thank you for your support!

Follow Us on Twitter!

:(