• @danielfgom
    link
    English
    201 year ago

    What’s the big deal? Google already do this to Android AND they also host Hackathons where they invite people to do this and reward those who do.

    Plus they have an ongoing bug bounty program so at anytime you can submit a bug/hack and get paid.

    Apple don’t have a history of doing this at all. This is literally the first time they are doing it because of the bad PR from Pegasus.

    It shouldn’t be applauded. They should be roasted for not having done this sooner

    • @[email protected]
      link
      fedilink
      English
      11 year ago

      I haven’t heard about google testing hardware based attacks on their chips, which I suppose could be caused by android running on a wide variety of chips instead of a few home-developed ones. Next to that Apple has had a bug bounty program for ages, that pays well and covers a wide range of attacks. Not hosting open hackathons has perhaps something to do with public brand image, but Apple shouldn’t be discredited regarding rewarding the findings of bugs and exploits.

      • @danielfgom
        link
        English
        61 year ago

        Not sure about their own chips but definitely hack the daylights out of Android.

        Apple has a bounty program but it doesn’t work. I’ve read multiple stories over the years of Devs who submitted show stopping bugs and never got anything back from Apple. And they take MONTHS to release a fix.

        The Google Security Team found a massive hole in iOS, reported it to Apple, and after months of waiting with no feedback or fix released, they published it openly. Only THEN did Apple suddenly acknowledge it and released a fix.

        Apple are the biggest hypocrites. They claim to be private and not collect data but literally everything you do on your phone they can see and collect. Everything in iCloud is on their servers. All your browsing history they can see in Safari.

        The only difference between them and Google is that they claim not to sell the data. But as we know Edward Snowden told us that the CIA/FBI etc have full access to all the servers of the Big Tech companies under the Patriot Act. They can decrypt and see your data anytime.

        So in other words not really private. None of them are.

        • @[email protected]
          link
          fedilink
          English
          21 year ago

          Sadly the same thing has been happening on the android side (a quick google search seems to confirm this). Possible exploits reported but not patched in a timely manner. In general I feel like the Apple bug bounty problem has been swift, although indeed failing from time to time to reward an original reporter. I have not been keeping a close eye on the android side but I imagine the same has been happening. Apple has started to offer e2e encryption on iCloud data blocking even CIA/FBI access. And next to that, seeing I’m based in Europe (and so my data should too) I don’t feel like the patriot act has any impact on me.

          • @danielfgom
            link
            English
            -11 year ago

            I’m assuming that Big Tech holds the encryption keys which they give the government access to in order to decrypt your data. The point of the Act is to allow law enforcement to be able to legibly access data in order to investigate possible terrorists.

            It wouldn’t be a very useful Act if they don’t hold the decryption keys. So they definitely do.

            And Snowden is still wanted, which means the info he leaked is accurate.

            I too am in the EU but I don’t trust any government. I’m sure they also can get the access from the US if they really want to. No one’s data is truly safe if you’re using Big Tech.

            • @[email protected]
              link
              fedilink
              English
              21 year ago

              Having someone else with the decryption keys is not how e2e works. E2E is a pretty solid and proven system, and I have yet to find a solid source about “big tech holding the keys”.

              • @danielfgom
                link
                English
                21 year ago

                Search for Project Prism

                Here’s one exhibit:

                "The National Security Agency has obtained direct access to the systems of Google, Facebook, Apple and other US internet giants, according to a top secret document obtained by the Guardian.

                The NSA access is part of a previously undisclosed program called Prism, which allows officials to collect material including search history, the content of emails, file transfers and live chats, the document says.

                The Guardian has verified the authenticity of the document, a 41-slide PowerPoint presentation – classified as top secret with no distribution to foreign allies – which was apparently used to train intelligence operatives on the capabilities of the program. The document claims “collection directly from the servers” of major US service providers."

                Source: https://www.theguardian.com/world/2013/jun/06/us-tech-giants-nsa-data

                That was back in 2013. I’m sure the tool is even more advanced now. This is why Snowden fled - he exposed this.

                • @[email protected]
                  link
                  fedilink
                  English
                  21 year ago

                  Yeah as the previous commenter said, e2e encryption just doesn’t allow anyone to access the data but the owner of the keys. E2E is prized because of this. There are two keys: public and private. If you and I are both using iMessage, you send a message to me that is encrypted on your device using your private key, and sent to my device using my public key. Only you and I can ever see those messages unless someone gets access to one of our phones.

                  Now, iCloud is backed up to apples servers. If you have iMessage backup enabled, it’s possible, and maybe even likely tbh, that Apple has access to recent messages. iMessage is also (potentially, but again in this case, I’d argue likely) susceptible to man-in-the-middle attacks. Because you need my public key for our communication to be decrypted, if you receive some else’s public key instead, they now have your messages and I don’t.

                  The DEA and FBI have both had documents leaked mentioning they can’t track or trace or unencrypt iMessage. The same is true for WhatsApp or any e2e messaging service.

                  Again, this is all contingent on not using iCloud backup. If you use iCloud backup, then the encryption keys used can be accessed with the proper authority. I assume (but haven’t looked into it) that Google is the same. If you don’t backup your e2e encrypted content, it cannot be decrypted without the private key only you have access to. Of course iCloud backup is enabled by default, so for the vast majority of Apple users, their messages and information are all available anyway so none of this matters.

                  In addition, iMessage uses a directory lookup to find the correct public key for your recipient. This information Apple does keep (I am unsure how long). What this means is that law enforcement (with a warrant) can see who and how often you are messaging. That alone is information we really don’t want people having.

                  So the moral is: if you don’t use backups for e2e encrypted communication, your content cannot be read externally. It’s just the way cryptography works.

                  This doesn’t mean that companies do not share information with law enforcement. There is a lot of unencrypted information Apple, Google, et al will share with government agencies when a warrant or subpoena is served. In addition to that, your phone provider will share information with them. In addition to that any SMS or MMS messages sent from any device will lack encryption and be easily discoverable.

                  Tl;dr: e2e encryption is secure, as long as you follow best practices and have an idea of how encryption works.

                  • @danielfgom
                    link
                    English
                    11 year ago

                    Just a thought: how can the public key unencrypt something that was encrypted with a private key? Wouldn’t the receiver need the same private key that was used to create the encryption.

                    Normally I’d encrypt a hard drive, for example, you’re asked for a password to decrypt it again. And it also generates a long recovery key. If I try to decrypt with the wrong password, it won’t decrypt.

                    Another thought: surely the sender must be sending the key to the receiver via a server (eg WhatsApp or iCloud)? If so what prevents that server from keeping a copy of the key and the message for future decryption by law enforcement?

                    Thanks