Options
Apple's refusal to help in unlocking terrorist's phone - RIGHT or WRONG?
heresitting
Posts: 2,743
Forum Member
✭✭✭
WRONG for me.
That sanctimonious outpouring from Tim Cook is beyond the pale. Hope their share price plummets.
[ don't know how to do a poll else I would have set up one]
That sanctimonious outpouring from Tim Cook is beyond the pale. Hope their share price plummets.
[ don't know how to do a poll else I would have set up one]
0
Comments
"For the children, for the children" NO. Giving up sensitive security code and a backdoor way in which no doubt the government will abuse and use to access innocent peoples devices just isn't worth the risk.
You do realise that whilst Apple are keeping your phone secure so your missus doesn't find out how much porn you are accessing they are simultaneously hovering up all that data for their own purposes?
I understand the argument, but anyone who thinks that Apple really care about our privacy are very naive. This is a calculated PR stunt, because unlocking the phone (or helping to) could put customers off, discredit their security claims and dent sales.
They are also flirting with a negative blowback...I doubt Apple want their image lumped in with Kalashnikov and the Toyota Hilux as "the first choice phone for the black flag waving men about town" ;-)
You know that for sure?
Yep, which is why Apple should comply.
Indeed. Or a double bluff to get bad people using Apple phones. They're unhackable. Risk Apple faces is an expansion of this-
https://en.wikipedia.org/wiki/Communications_Assistance_for_Law_Enforcement_Act
to cover mobile networks and devices. And as you say, Apple has scant regard for people's privacy, as evidenced in their implementations and T&Cs. Plus modifying security features may make it easier for owners to jailbreak iPhones.
I'd say it's obvious considering governments stance on civilian encryption. Giving the government access to sensitive code will allow them to implement ways in which they could circumvent all phones security which again is not on! Apple should be commended for telling them no.
Is it that making such software it would put there phones at risk, how and why?
Their software writers would had to write software to protect the phones anyway.
If they are saying they wish to protect their customers personal data, that's a bit bizarre as they are ones accessing that private data for commercial reasons. What's all this protecting personal data because it's on a phone or computer, did safemakers refuse to open safes for the authorities because it contained personal data?
When data was on paper kept at home...did magistrates refuse to issue search warrants to protect peoples data?
Apple are saying they would have to create a way to unlock the system and if they did that then foreign countries such as China could state they want this technology in any phones sold in the country, they also say it opens up their systems to hackers.
They say they have built their company around security and privacy, they are also saying they would effectively be working for the FBI by being forced to create a way to hack the phone.
It seems to me like the FBI is using version 2 of "won't someone think of the children" - this time it's "won't someone think about stopping terrorism" - even though the risk is already extremely low and decrypting this phone is not likely to lower it further.
Then, they can get Apple to install the "FBiOS" (as some call it) onto every iPhone and iPad they find, regardless of crime, because it'll be much easier to get a warrant when it's been proven that it's possible. I can imagine palletloads of iDevices heading to Cupertino to be backdoored.
If it's not the case already, I bet a future iPhone has the passcode stuff implemented in ROM, so there's no way Apple can get around it
As I understand it, Apple have a phone which cannot be hacked. If they then create a back door into the phone which currently does not exist then they make every one of their phones open to hacking and not just by the FBI. Doesn't matter how many locks and bolts they put on the back door it has made an inpregnable system open to hacking.
Apple have a totally "safe" product but creating new software to look at one phone will make it possible for anyone with the expertise to hack any of the new iPhone.
On the other hand should Apple provide a product that in essence can be used for terror without having some sort of way of being able to access it if required.
Are Apple responsible for the people who use and how they use it?
Phones are tracked on a regular basis for example if someone is suicidal the police can track the phone, a policeman friend of mine arrested an 'alleged' serial killer using mobile phone records e.g. cell activity etc, non of that required access to encrypted information, Apple are right to stand their ground.
Not so much of an excuse more of a test case. I would say that Apple have picked the wrong case to make a stand over - public opinion in the US could easily swing against them if they continue to refuse to assist an investigation into Islamic Terrorism on US soil - the PR consequences could cost millions.
Only if they distribute this new software, which nobody is asking for.