By continuing to use the site or forum, you agree to the use of cookies, find out more by reading our GDPR policy

The modular and most adaptive version of Windows 10 is currently called ‘Windows 10X’ and it’ll be arriving on traditional single-screen laptops in the first half of 2021. Windows 10’s modular version was first announced in October 2019 and Microsoft originally said that the Surface Neo would be the first device to run the new OS. However, everything has changed after Microsoft started preparing Windows 10X for single-screen devices in an effort to meet the current needs of the customers. Microsoft has also removed the Surface Neo listing to clarify it’s not coming this holiday season. At the moment, we’re not sure if Microsoft will publish the beta builds of Windows 10X to the testers in the Insider, which raises fresh questions and concerns about exactly how the OS might work. Windows 10X is also known as Windows Lite internally and it’s based on Windows Core OS, which modularizes Windows Shell and other components. This new operating system is designed to run on both single-screen and dual-screen form factors, and it’s also modern without legacy components. In addition, Windows 10X comes with a new user interface that ditches live tiles support for icons and it also allows Windows Update to happen seamlessly in the background. According to sources, Microsoft is currently planning to deliver Windows 10X sometime in the first half of 2021 without native support for Win32 apps. The first single-screen Windows 10X PCs are also set to arrive in the second quarter or Spring of 2021. To make room for Windows 10X launch, Microsoft also appears to be considering changes to its Windows 10 upgrade cycle. Going forward, Windows 10 will receive only one feature update per year and next year’s feature update is expected to arrive after Windows 10X launches in the market. After the launch of Windows 10X in Q2 2021, Microsoft will begin rolling out the first feature update for Windows 10. In the spring of 2022, we’ll see the first big Windows 10X feature update that will add support for dual-screen hardware, such as Surface Neo and Lenovo ThinkPad Fold. Unfortunately, Microsoft has reportedly removed the virtualization technology from the internal builds of Windows 10X. This would have allowed Win32 apps (desktop or classic apps) to run smoothly in a container. Microsoft is not satisfied with the performance of Win32 apps on Windows 10X due to limitations. For example, some Win32 apps are struggling to access the native features available outside the container, which includes screen sharing and alerts when apps are minimized to the taskbar. This is the opposite of the Windows 10X ethos, which is supposed to offer both performance and compatibility at the same time. As a result, Windows 10X internal builds have dropped support for Win32 apps. You can only run UWP and web apps natively, which would turn Windows 10X into a proper lightweight OS for Chromebook-like devices. Microsoft will allow early adopters to stream Win32 apps via a web service, which works only when you have an internet connection. While the plans are always subject to change, Microsoft has internally decided not to move forward with the dual-screen model for another year. For more on this visit OUR FORUM often.

FEDERAL AGENTS from the Department of Homeland Security and the Justice Department used “a sophisticated cell phone cloning attack—the details of which remain classified—to intercept protesters’ phone communications” in Portland this summer, Ken Klippenstein reported this week in The Nation. Put aside for the moment that, if the report is true, federal agents conducted sophisticated electronic surveillance against American protesters, an alarming breach of constitutional rights. Do ordinary people have any hope of defending their privacy and freedom of assembly against threats like this? Without more details, it’s hard to be entirely sure what type of surveillance was used, but The Nation’s mention of “cell phone cloning” makes me think it was a SIM cloning attack. This involves duplicating a small chip used by virtually every cellphone to link itself to its owner’s phone number and account; this small chip is the subscriber identity module, more commonly known as SIM.  SIM cards contain a secret encryption key that is used to encrypt data between the phone and cellphone towers. They’re designed so that this key can be used (like when you receive a text or call someone) but so the key itself can’t be extracted. But it’s still possible to extract the key from the SIM card, by cracking it. Older SIM cards used a weaker encryption algorithm and could be cracked quickly and easily, but newer SIM cards use stronger encryption and might take days or significantly longer to crack. It’s possible that this is why the details of the type of surveillance used in Portland “remain classified.” Do federal agencies know of a way to quickly extract encryption keys from SIM cards? (On the other hand, it’s also possible that “cell phone cloning” doesn’t describe SIM cloning at all but something else instead, like extracting files from the phone itself instead of data from the SIM card.) Assuming the feds were able to extract the encryption key from their target’s SIM card, they could give the phone back to their target and then spy on all their target’s SMS text messages and voice calls going forward. To do this, they would have to be physically close to their target, monitoring the radio waves for traffic between their target’s phone and a cell tower. When they see it, they can decrypt this traffic using the key they stole from the SIM card. This would also fit with what the anonymous former intelligence officials told The Nation; they said the surveillance was part of a “Low-Level Voice Intercept” operation, a military term describing audio surveillance by monitoring radio waves. Even if law enforcement agencies don’t clone a target’s SIM card, they could gather quite a bit of information after temporarily confiscating the target’s phone. They could power off the phone, pop out the SIM card, put it in a separate phone, and then power that phone on. If someone sends the target an SMS message (or texts a group that the target is in), the feds’ phone would receive that message instead of the target’s phone. And if someone called the target’s phone number, the feds’ phone would ring instead. They could also hack their target’s online accounts, so long as those accounts support resetting the password using a phone number. But, in order to remain stealthy, they would need to power off their phone, put the SIM card back in their target’s phone, and power that phone on again before returning it, which would restore the original phone’s access to the target’s phone number, and the feds would lose access. Read this entire posting on OUR FORUM.

Researchers have uncovered a threat group launching surveillance campaigns that target victims’ personal device data, browser credentials, and Telegram messaging application files. One notable tool in the group’s arsenal is an Android malware that collects all two-factor authentication (2FA) security codes sent to devices, sniffs out Telegram credentials, and launches Google account phishing attacks. Researchers found the threat group, dubbed Rampant Kitten, has targeted Iranian entities with surveillance campaigns for at least six years. It specifically targets Iranian minorities and anti-regime organizations, including the Association of Families of Camp Ashraf and Liberty Residents (AFALR); and the Azerbaijan National Resistance Organization. The threat group has relied on a wide array of tools for carrying out their attacks, including four Windows info-stealer variants used for pilfering Telegram and KeePass account information; phishing pages that impersonate Telegram to steal passwords; and the aforementioned Android backdoor that extracts 2FA codes from SMS messages and records the phone’s voice surroundings. “Following the tracks of this attack revealed a large-scale operation that has largely managed to remain under the radar for at least six years,” said researchers with Check Point Research, in a Friday analysis. “According to the evidence we gathered, the threat actors, who appear to be operating from Iran, take advantage of multiple attack vectors to spy on their victims, attacking victims’ personal computers and mobile devices.” The Attacks Researchers first discovered Rampant Kitten’s campaign through a document, the title of which translates to “The Regime Fears the Spread of the Revolutionary Cannons.docx.” It’s unclear how this document is spread (via spear-phishing or otherwise), but it purports to describe the ongoing struggle between the Iranian regime and the Revolutionary Cannons, an anti-regime, Mujahedin-e Khalq movement. The document when opened loads a document template from a remote server (afalr-sharepoint[.]com), which impersonates a website for a non-profit that aids Iranian dissidents. It then downloads malicious macro code, which executes a batch script to download and execute a next-stage payload. This payload then checks if the popular Telegram messenger service is installed on the victims’ system. If so, it extracts three executables from its resources. These executables include an information stealer, which lifts Telegram files from the victim’s computer, steals information from the KeePass password-management application, uploads any file it can find which ends with a set of pre-defined extensions, and logs clipboard data and takes desktop screenshots. Researchers were able to track multiple variants of this payload dating back to 2014. These include the TelB (used in June and July 2020) and TelAndExt variants (May 2019 to February 2020), which focus on Telegram; a Python info stealer (February 2018 to January 2020) that is focused on stealing data from Telegram, Chrome, Firefox and Edge; and a HookInjEx variant (December 2014 to May 2020), an info stealer that targets browsers, device audio, keylogging and clipboard data. During their investigation, researchers also uncovered a malicious Android application tied to the same threat actors. The application was purporting to be a service to help Persian speakers in Sweden get their driver’s license. Instead, once victims download the application, the backdoor steals their SMS messages and bypasses 2FA by forwarding all SMS messages containing 2FA codes to an attacker-controlled phone number. “One of the unique functionalities in this malicious application is forwarding any SMS starting with the prefix G- (The prefix of Google two-factor authentication codes) to a phone number that it receives from the C2 server,” said researchers. “Furthermore, all incoming SMS messages from Telegram, and other social network apps, are also automatically sent to the attackers’ phone number.” Of note, the application also launches a phishing attack targeting victims’ Google account (Gmail) credentials. The user is presented with a legitimate Google login page, inside Android’s WebView. In reality, attackers have used Android’s JavascriptInterface to steal typed-in credentials, as well as a timer that periodically retrieves the information from the username and password input fields. We have more of this posted on OUR FORUM.

A newly discovered technique by a researcher shows how Google's App Engine domains can be abused to deliver phishing and malware while remaining undetected by leading enterprise security products. Google App Engine is a cloud-based service platform for developing and hosting web apps on Google's servers. While reports of phishing campaigns leveraging enterprise cloud domains are nothing new, what makes Google App Engine infrastructure risky in how the subdomains get generated and paths are routed. Typically scammers use cloud services to create a malicious app that gets assigned a subdomain. They then host phishing pages there. Or they may use the app as a command-and-control (C2) server to deliver malware payload. But the URL structures are usually generated in a manner that makes them easy to monitor and block using enterprise security products, should there be a need. Therefore, a cybersecurity professional could block traffic to and from this particular app by simply blocking requests to and from this subdomain. This wouldn't prevent communication with the rest of the Microsoft Azure apps that use other subdomains. It gets a bit more complicated, however, in the case of Google App Engine. Security researcher Marcel Afrahim demonstrated an intended design of Google App Engine's subdomain generator, which can be abused to use the app infrastructure for malicious purposes, all while remaining undetected. A subdomain, in this case, does not only represent an app, it represents an app's version, the service name, project ID, and region ID fields. But the most important point to note here is, if any of those fields are incorrect, Google App Engine won't show a 404 Not Found page, but instead show the app's "default" page (a concept referred to as soft routing). "Requests are received by any version that is configured for traffic in the targeted service. If the service that you are targeting does not exist, the request gets Soft Routed," states Afrahim, adding: "If a request matches the portion of the hostname, but includes a service, version, or instance name that does not exist, then the request is routed to the default service, which is essentially your default hostname of the app." Essentially, this means there are a lot of permutations of subdomains to get to the attacker's malicious app. As long as every subdomain has a valid "project_ID" field, invalid variations of other fields can be used at the attacker's discretion to generate a long list of subdomains, which all lead to the same app. The fact that a single malicious app is now represented by multiple permutations of its subdomains makes it hard for sysadmins and security professionals to block malicious activity. But further, to a technologically unsavvy user, all of these subdomains would appear to be a "secure site." After all, the domain and all its subdomains come with the seal of "Google Trust Services" in their SSL certificates. Even further, most enterprise security solutions such as Symantec WebPulse web filter automatically allow traffic to trusted category sites. And Google's domain, due to its reputation and legitimate corporate use cases, earns an "Office/Business Applications" tag, skipping the scrutiny of web proxies. This complete article is posted on OUR FORUM with much more information.

Intel's slow trickle of information on its Tiger Lake processors recently turned into a veritable flood as the company shared information about its first salvo of 10nm SuperFin chips, but one detail was missing: Any official disclosures of chips with more than four cores. That changed in a decidedly low-key way, as a blog post from Intel fellow Boyd Phelps on Medium reveals that the company will introduce eight-core models soon, saying: "We also added a 3MB non-inclusive last-level-cache (LLC) per core slice. A single core workload has access to 12MB of LLC in the 4-core die or up to 24MB in the 8-core die configuration (more detail on 8-core products at a later date)." Intel claims that it's four-core Tiger Lake models, by virtue of their 10nm SuperFin Process, Willow Cove Cores, and Iris XE graphics can already beat AMD's eight-core Renoir chips in some performance benchmarks. If Intel's performance projections for its quad-core models are accurate, the eight-core Tiger Lake models could prove to be exceedingly competitive against AMD's existing Ryzen Mobile 'Renoir' lineup, possibly even wresting away the lead in threaded applications. We've yet to see independent third-party verification of the quad-core Tiger Lake chips in reviews, but AMD's upcoming Zen 3 "Cezanne" APUs are now extremely important as AMD looks to keep its performance advantage in the laptop market despite the looming eight-core Tiger Lake models. The current dual- and quad-core Tiger Lake chips address only the 7 to 28W segment, while larger eight-core Tiger Lake-H processors would obviously tackle the upper echelons of the performance market, possibly stretching up to 45W models (~65W peak) for H-series Core i9 and i7 models. We won't go into Tiger Lake's full technical details, we have all of those resources in one place here, but Intel's plans for eight-core Tiger Lake models aren't entirely surprising. Intel's current 10th-gen lineup includes 10nm Ice Lake processors that address the iGPU gaming market with up to four cores, while the 14nm Comet Lake processors slot in for high-performance productivity workloads. However, Intel told us during its Tiger Lake briefings that all of its future laptop chips will come with the 10nm SuperFin (or better) process, meaning the company won't have a split product stack for its 11th-gen lineup. Much of Intel's previous limitations on its Ice Lake models stemmed from the low clock frequencies and poor yields, both of which conspired to limit performance and core counts - Intel's best 10nm efforts thus far have resulted in quad-core chips for laptops. Intel's new 10nm SuperFin process has corrected the clock speed issues, we see up to a 700 MHz increase to base and boost frequencies, and the emergence of eight core models imply that defect rates are lower, and thus yields are up, allowing Intel to punch out 10nm laptop chips with up to eight cores. Intel has no plans to bring Tiger Lake to its lineup of desktop chips, but we have already seen the first new Tiger Lake NUCs emerge from ASRock. Naturally, eight-core Tiger Lake models will also work their way into the NUC lineups. Given their pairing with the Xe graphics engine, they could prove to pack a decent performance punch for compact desktop PCs. Stay abreast on this and other news from Intel by visiting OUR FORUM.

Earlier this summer, marine specialists reeled up a shipping-container-size datacenter coated in algae, barnacles, and sea anemones from the seafloor off Scotland’s Orkney Islands. The retrieval launched the final phase of a years-long effort that proved the concept of underwater datacenters is feasible, as well as logistically, environmentally, and economically practical. Microsoft’s Project Natick team deployed the Northern Isles datacenter 117 feet deep to the seafloor in spring 2018. For the next two years, team members tested and monitored the performance and reliability of the datacenter’s servers. The team hypothesized that a sealed container on the ocean floor could provide ways to improve the overall reliability of data centers. On land, corrosion from oxygen and humidity, temperature fluctuations, and bumps and jostles from people who replace broken components are all variables that can contribute to equipment failure. The Northern Isles deployment confirmed their hypothesis, which could have implications for data centers on land. Lessons learned from Project Natick also are informing Microsoft’s datacenter sustainability strategy around energy, waste, and water, said Ben Cutler, a project manager in Microsoft’s Special Projects research group who leads Project Natick. What’s more, he added, the proven reliability of underwater datacenters has prompted discussions with a Microsoft team in Azure that’s looking to serve customers who need to deploy and operate tactical and critical datacenters anywhere in the world. “We are populating the globe with edge devices, large and small,” said William Chappell, vice president of mission systems for Azure. “To learn how to make data centers reliable enough not to need human touch is a dream of ours.” The underwater datacenter concept splashed onto the scene at Microsoft in 2014 during ThinkWeek, an event that gathers employees to share out-of-the-box ideas. The concept was considered a potential way to provide lightning-quick cloud services to coastal populations and save energy. More than half the world’s population lives within 120 miles of the coast. By putting datacenters underwater near coastal cities, data would have a short distance to travel, leading to fast and smooth web surfing, video streaming, and game playing. The consistently cool subsurface seas also allow for energy-efficient datacenter designs. For example, they can leverage heat-exchange plumbing such as that found on submarines. Microsoft’s Project Natick team proved the underwater datacenter concept was feasible during a 105-day deployment in the Pacific Ocean in 2015. Phase II of the project included contracting with marine specialists in logistics, shipbuilding, and renewable energy to show that the concept is also practical. “We are now at the point of trying to harness what we have done as opposed to feeling the need to go and prove out some more,” Cutler said. “We have done what we need to do. Natick is a key building block for the company to use if it is appropriate.” We have pictures, videos, and more posted on OUR Forum.