close
    • chevron_right

      Twitter posts the code it claims determines which tweets people see, and why

      news.movim.eu / ArsTechnica · Friday, 31 March, 2023 - 22:24

    Section of Twitter's source code, displayed at an angle

    Enlarge / Twitter has posted what it states is the code used by its algorithm to recommend tweets to its users.

    Twitter has made good on one of CEO Elon Musk's many promises , posting on a Friday afternoon what it claims is the code for its tweet recommendation algorithm on GitHub .

    The code, posted under a GNU Affero General Public License v3.0 , contains numerous insights as to what factors make a tweet more or less likely to show up in users' timelines.

    In a blog post accompanying the code release , Twitter's engineering team (under no particular byline) notes that the system for determining which "top Tweets that ultimately show up on your device's For You timeline" is "composed of many interconnected services and jobs." Each time a Twitter home screen is refreshed, Twitter pulls "the best 1,500 Tweets from a pool of hundreds of millions," the post states.

    Read 11 remaining paragraphs | Comments

    • chevron_right

      SIKE Broken

      news.movim.eu / Schneier · Wednesday, 3 August, 2022 - 09:03

    SIKE is one of the new algorithms that NIST recently added to the post-quantum cryptography competition.

    It was just broken , really badly.

    We present an efficient key recovery attack on the Supersingular Isogeny Diffie­-Hellman protocol (SIDH), based on a “glue-and-split” theorem due to Kani. Our attack exploits the existence of a small non-scalar endomorphism on the starting curve, and it also relies on the auxiliary torsion point information that Alice and Bob share during the protocol. Our Magma implementation breaks the instantiation SIKEp434, which aims at security level 1 of the Post-Quantum Cryptography standardization process currently ran by NIST, in about one hour on a single core.

    News article .

    • chevron_right

      Mathias Poujol-Rost ✅ · Sunday, 6 February, 2022 - 17:14

      Contact publication

    FediTips (@feditips@mstdn.social)
    • FediTips (@feditips@mstdn.social)

      Mastodon and most other Fediverse platforms display posts chronologically, without any algorithms. On your feed you will see all the posts from all the people you follow, in the order they posted them. Because of this, if you post something important you may want to boost it again later in the day so that people in different time zones have a better chance of seeing it. #MastoTips #FediTips #Mastodon #Fediverse #TimeZones

    • favorite

      2 Like

      jorgeluis, Timothée Jaussoin

    • chevron_right

      D-Wave releases its next-generation quantum annealing chip

      John Timmer · news.movim.eu / ArsTechnica · Tuesday, 29 September, 2020 - 18:13 · 1 minute

    Image of a chip surrounded by complicated support hardware.

    Enlarge

    Today, quantum computing company D-Wave is announcing the availability of its next-generation quantum annealer, a specialized processor that uses quantum effects to solve optimization and minimization problems. The hardware itself isn't much of a surprise—D-Wave was discussing its details months ago —but D-Wave talked with Ars about the challenges of building a chip with over a million individual quantum devices. And the company is coupling the hardware's release to the availability of a new software stack that functions a bit like middleware between the quantum hardware and classical computers.

    Quantum annealing

    Quantum computers being built by companies like Google and IBM are general purpose, gate-based machines. They can solve any problem and should show a vast acceleration for specific classes of problems. Or they will, as soon as the gate count gets high enough. Right now, these quantum computers are limited to a few dozen gates and have no error correction. Bringing them up to the scale needed presents a series of difficult technical challenges.

    D-Wave's machine is not general-purpose; it's technically a quantum annealer, not a quantum computer. It performs calculations that find low-energy states for different configurations of the hardware's quantum devices. As such, it will only work if a computing problem can be translated into an energy-minimization problem in one of the chip's possible configurations. That's not as limiting as it might sound, since many forms of optimization can be translated to an energy minimization problem, including things like complicated scheduling issues and protein structures.

    Read 22 remaining paragraphs | Comments

    index?i=WBWYyxMn2Kw:7SblLUC8gZY:V_sGLiPBpWUindex?i=WBWYyxMn2Kw:7SblLUC8gZY:F7zBnMyn0Loindex?d=qj6IDK7rITsindex?d=yIl2AUoC8zA