(2023-07-14, 08:09 PM)Sciborg_S_Patel Wrote: [ -> ]My reasoning for new hardware is the kinds of programming languages that I think could make a real AI driver - Idris for example - seem to be too slow for the kind of hard real time reactions driving requires.
As for how close we are, the shady way this tech is pushed onto the public and the complaints/issues are reflected in China as well. No one seems to be making enough headway and investor confidence has dropped.
Maybe someone will have a breakthrough, but IMO the fundamental nature of machine learning leads me to suspect it's not possible. I'd also point out that those who would hack these machine learning drivers have not really stepped up yet but that too is inevitable, and machine learning - which as produced these fatalities and accidents already - seems quite vulnerable in this regard.
I fact checked Idris on the Internet. There’s nothing about this programming language that seperate it from any other programming language in terms of hardware requirements. The only paradigm breaking kind of computation is quantum computing which requires yet to be invented hardware.
(2023-07-15, 12:13 PM)sbu Wrote: [ -> ]I fact checked Idris on the Internet. There’s nothing about this programming language that seperate it from any other programming language in terms of hardware requirements. The only paradigm breaking kind of computation is quantum computing which requires yet to be invented hardware.
It's slower than what you'd want for hard real time systems. That was my only point.
(2023-07-15, 01:42 PM)Sciborg_S_Patel Wrote: [ -> ]It's slower than what you'd want for hard real time systems. That was my only point.
This may be. But the software isn’t ready yet either. I still don’t understand that self-driving can be legal anywhere either.
Cruise blames Outside Lands for driverless car traffic fiasco in San Francisco
Laya Neelakandan, Ricardo Cano
Quote:UPDATE: Supervisor Aaron Peskin said Sunday that government agencies would ask the city attorney to file a petition requesting that the state revisit last week’s key approval expanding robotaxi service.
As many as 10 Cruise driverless cars stopped working in San Francisco’s North Beach on Friday night, causing traffic to back up and leaving some questioning the decision of state regulators a day earlier to approve the expanded use of robotaxis in the city.
The autonomous vehicles appeared to be stopped in the middle of Grant Avenue, according to social media posts, with hazard lights on, blocking other cars from moving.
In a response to the incident, Cruise said the backup was caused by “wireless connectivity issues” that immobilized the driverless cars. San Francisco police confirmed that the cell connectivity issues were caused by the large number of people at the nearby Outside Lands music festival overtaxing the system.
“We are actively investigating and working on solutions to prevent this from happening again and apologize to those impacted,” Cruise said in a statement.
I guess nobody had thought to include pictures of wet concrete in the images used to train the AI!
David
Well team: I purchased a 2023 Model Y a few weeks ago. Have to report that so far I simply love the car.
I did NOT purchase the "Full Self Driving" module but the car comes standard with a quasi-version of autopilot. Basically, it has adaptive cruise control combined with auto-steering lane maintenance. Its really wonderful on longer highway drives (which I do with some regularity). Someone put it this way to me: "Teslas are technology with a car wrapped around it. Other automakers make cars with technology wrapped around them." That resonates with me.
I did get a loaner after having a couple of minor delivery items addressed. The loaner (a Model 3) DID have FSD. So, I told the car to drive me home. It was pretty remarkable. Changed lanes, stopped for lights, etc. It basically did drive me home. However, and its a BIG however, the car did get confused on a left turn (across traffic here in the states) at a traffic light. The light went yellow and then red while the car was a teeny bit extended into the intersection. That was well and good. However, a couple seconds later after cross traffic was fully active it decided NOW was the time to make the left turn. Needless to say I retook control of the vehicle and stopped its movement. This would have clearly been an accident.
So, to Sci's point..... Tesla's FSD isn't quite ready for prime time. That said it sure felt 'close' to me but I'll concede the proverbial last mile is probably the hardest part by a big factor.
(2023-08-22, 03:28 AM)Silence Wrote: [ -> ]So, to Sci's point..... Tesla's FSD isn't quite ready for prime time. That said it sure felt 'close' to me but I'll concede the proverbial last mile is probably the hardest part by a big factor.
I'd say the challenge is that very few people - if anyone - can give you a clear measure of how close self-driving is to completion.
Personally I'd say it's like Russian Roulette, and it's a good thing you were paying attention and took control.
(2023-08-22, 03:28 AM)Silence Wrote: [ -> ]Well team: I purchased a 2023 Model Y a few weeks ago. Have to report that so far I simply love the car.
I did NOT purchase the "Full Self Driving" module but the car comes standard with a quasi-version of autopilot. Basically, it has adaptive cruise control combined with auto-steering lane maintenance. Its really wonderful on longer highway drives (which I do with some regularity). Someone put it this way to me: "Teslas are technology with a car wrapped around it. Other automakers make cars with technology wrapped around them." That resonates with me.
I did get a loaner after having a couple of minor delivery items addressed. The loaner (a Model 3) DID have FSD. So, I told the car to drive me home. It was pretty remarkable. Changed lanes, stopped for lights, etc. It basically did drive me home. However, and its a BIG however, the car did get confused on a left turn (across traffic here in the states) at a traffic light. The light went yellow and then red while the car was a teeny bit extended into the intersection. That was well and good. However, a couple seconds later after cross traffic was fully active it decided NOW was the time to make the left turn. Needless to say I retook control of the vehicle and stopped its movement. This would have clearly been an accident.
So, to Sci's point..... Tesla's FSD isn't quite ready for prime time. That said it sure felt 'close' to me but I'll concede the proverbial last mile is probably the hardest part by a big factor.
Can I just clarify something? You made ONE journey in the self driving car, and you had to take over to prevent a bad crash!
I think that says it all.
Did you report it to the company?
David