id
stringlengths
3
8
url
stringlengths
32
207
title
stringlengths
1
114
text
stringlengths
93
492k
1129601
https://en.wikipedia.org/wiki/Ad%20blocking
Ad blocking
Ad blocking or ad filtering is a software capability for blocking or altering online advertising in a web browser, an application or a network. This may be done using browser extensions or other methods. Technologies and native countermeasures Online advertising exists in a variety of forms, including web banners, pictures, animations, embedded audio and video, text, or pop-up windows, and can even employ audio and video autoplay. Many browsers offer some ways to remove or alter advertisements: either by targeting technologies that are used to deliver ads (such as embedded content delivered through browser plug-ins or via HTML5), targeting URLs that are the source of ads, or targeting behaviors characteristic to ads (such as the use of HTML5 AutoPlay of both audio and video). Prevalence Use of mobile and desktop ad blocking software designed to remove traditional advertising grew by 41% worldwide and by 48% in the U.S. between Q2 2014 and Q2 2015. As of Q2 2015, 45 million Americans were using ad blockers. In a survey research study released Q2 2016, Met Facts reported 72 million Americans, 12.8 million adults in the UK, and 13.2 million adults in France were using ad blockers on their PCs, smartphones, or tablet computers. In March 2016, the Internet Advertising Bureau reported that UK adblocking was already at 22% among people over 18 years old. As of 2021, 27% of US Internet users used ad blocking software, with continued increasing trend since 2014. Among technical audiences the rate of blocking reaches 58% as of 2021. Reasons for blocking ads There are various fundamental reasons why one would want to use ad-blocking: Protecting privacy Reduces the number of HTTP cookies, browser fingerprinting and other aggressive behavioral tracking techniques Protecting from malvertising Any intrusive actions from the ads, including but not limited to: drive-by downloads, invisible overlay click areas (such as a regular link that opens an unexpected external website), opening in a new tab, popups and auto-redirects. Saving bandwidth (and by extension, money) On most websites user tracking and advertising code alone constitutes the majority of the downloadable content, thus significantly inflating the amount of data downloaded by the user ("web bloat") Better user experience Some ads cover the text making it partly illegible, making the site unusable Less cluttered pages Faster page loading times Fewer distractions Accessibility reasons The motion in some ads is nauseating for some users Save battery on mobile devices or laptops Prevent undesirable websites from making ad revenue out of the user's visit Publishers and their representative trade bodies, on the other hand, argue that web ads provide revenue to website owners, which enable the website owners to create or otherwise purchase content for the website. Publishers state that the prevalent use of ad blocking software and devices could adversely affect website owner revenue and thus, in turn, lower the availability of free content on websites. Benefits For users, the benefits of ad blocking software include quicker loading and cleaner looking web pages with fewer distractions, lower resource waste (bandwidth, CPU, memory, etc.), and privacy benefits gained through the exclusion of the tracking and profiling systems of ad delivery platforms. Blocking ads can also save substantial amounts of electrical energy and lower users' power bills, and additional energy savings can also be expected at the grid level because fewer data packets need to be transmitted between the user's machine and the website server. User experience Ad blocking software may have other benefits to users' quality of life, as it decreases Internet users' exposure to advertising and marketing industries, which promote the purchase of numerous consumer products and services that are potentially harmful or unhealthy and on creating the urge to buy immediately. The average person sees more than 5000 advertisements daily, many of which are from online sources. Ads may promise viewers that their lives will be improved by purchasing the item that is being promoted (e.g., fast food, soft drinks, candy, expensive consumer electronics) or encourages users to get into debt or gamble. Additionally, if Internet users buy all of these items, the packaging and the containers (in the case of candy and soda pop) end up being disposed of, leading to negative environmental impacts of waste disposal. Advertisements are very carefully crafted to target weaknesses in human psychology; as such, a reduction in exposure to advertisements could be beneficial for users' quality of life. Unwanted advertising can also harm the advertisers themselves if users become annoyed by the ads. Irritated users might make a conscious effort to avoid the goods and services of firms which are using annoying "pop-up" ads which block the Web content the user is trying to view. For users not interested in making purchases, the blocking of ads can also save time. Any ad that appears on a website exerts a toll on the user's "attention budget" since each ad enters the user's field of view and must either be consciously ignored or closed, or dealt with in some other way. A user who is strongly focused on reading solely the content that they are seeking likely has no desire to be diverted by advertisements that seek to sell unneeded or unwanted goods and services. In contrast, users who are actively seeking items to purchase, might appreciate advertising, in particular targeted ads. Security Another important aspect is improving security; online advertising subjects users to a higher risk of infecting their devices with computer viruses than surfing pornography websites. In a high-profile case, malware was distributed through advertisements provided to YouTube by a malicious customer of Google's Doubleclick. In August 2015, a 0-day exploit in the Firefox browser was discovered in an advertisement on a website. When Forbes required users to disable ad blocking before viewing their website, those users were immediately served with pop-under malware. The Australian Signals Directorate recommends individuals and organizations block advertisements to improve their information security posture and mitigate potential malvertising attacks and machine compromise. The information security firm Webroot also notes employing ad blockers provide effective countermeasures against malversating campaigns for less technically-sophisticated computer users. Monetary Ad blocking can also save money for the user. If a user's personal time is worth one dollar per minute, and if unsolicited advertising adds an extra minute to the time that the user requires for reading the webpage (i.e. the user must manually identify the ads as ads, and then click to close them, or use other techniques to either deal with them, all of which tax the user's intellectual focus in some way), then the user has effectively lost one dollar of time in order to deal with ads that might generate a few fractional pennies of display-ad revenue for the website owner. The problem of lost time can rapidly spiral out of control if malware accompanies the ads. Ad blocking also reduces page load time and saves bandwidth for the users. Users who pay for total transferred bandwidth ("capped" or pay-for-usage connections) including most mobile users worldwide have a direct financial benefit from filtering an ad before it is loaded. Using an ad blocker is a common method of improving internet speeds. Analysis of the 200 most popular news sites (as ranked by Alexa) in 2015 showed that Mozilla Firefox Tracking Protection lead to 39% reduction in data usage and 44% median reduction in page load time. According to research performed by The New York Times, ad blockers reduced data consumption and sped upload time by more than half on 50 news sites, including their own. Journalists concluded that "visiting the home page of Boston.com (the site with most ad data in the study) every day for a month would cost the equivalent of about $9.50 in data usage just for the ads". It is a known problem with most web browsers, including Firefox, that restoring sessions often plays multiple embedded ads at once. However, this annoyance can easily be averted simply by setting the web browser to clear all cookies and browsing-history information each time the browser software is closed. Another preventive option is to use a script blocker, which enables the user to disable all scripts and then to selectively re-enable certain scripts as desired, in order to determine the role of each script. The user thus can very quickly learn which scripts are truly necessary (from the standpoint of webpage functionality) and consequently which sources of scripts are undesirable, and this insight is helpful in visiting other websites in general. Thus by precisely controlling which scripts are run in each webpage viewed, the user retains full control over what happens on their computer CPU and computer screen. Methods One method of filtering is simply to block (or prevent auto play of) Flash animation or image loading or Microsoft Windows audio and video files. This can be done in most browsers easily and also improves security and privacy. This crude technological method is refined by numerous browser extensions. Every web browser handles this task differently, but, in general, one alters the options, preferences or application extensions to filter specific media types. An additional add-on is usually required to differentiate between ads and non-ads using the same technology, or between wanted and unwanted ads or behaviors. The more advanced ad-blocking filter software allows fine-grained control of advertisements through features such as blacklists, whitelists, and regular expression filters. Certain security features also have the effect of disabling some ads. Some antivirus software can act as an ad blocker. Filtering by intermediaries such as ISP providers or national governments is increasingly common. Browser integration As of 2015, many web browsers block unsolicited pop-up ads automatically. Current versions of Konqueror, Microsoft Edge, and Firefox also include content filtering support out-of-the-box. Content filtering can be added to Firefox, Chromium-based browsers, Opera, Safari, and other browsers with extensions such as AdBlock, Adblock Plus, and uBlock Origin, and a number of sources provide regularly updated filter lists. AdBlock Plus is included in the freeware browser Maxthon from the People's Republic of China by default. Another method for filtering advertisements uses Cascading Style Sheets (CSS) rules to hide specific HTML and XHTML elements. In January 2016, Brave, a free, ad-blocking browser for Mac, PC, Android, and iOS devices was launched. Brave users can optionally enable Brave's own ad network to earn Basic Attention Tokens (BATs), a type of cryptocurrency, which can be sent as micro-payments to publishers. At the beginning of 2018, Google confirmed that the built-in ad blocker for the Chrome/Chromium browsers would go live on 15 February: this ad blocker only blocks certain ads as specified by the Better Ads Standard (defined by the Coalition for Better Ads, in which Google itself is a board member). This built-in ad blocking mechanism is disputed because it could unfairly benefit Google's advertising itself. In 2019, both Apple and Google began to make changes to their web browsers' extension systems which encourage the use of declarative content blocking using pre-determined filters processed by the web browser, rather than filters processed at runtime by the extension. Both vendors have imposed limits on the number of entries that may be included in these lists, which have led to (especially in the case of Chrome) allegations that these changes are being made to inhibit the effectiveness of ad blockers. External programs A number of external software applications offer ad filtering as a primary or additional feature. A traditional solution is to customize an HTTP proxy (or web proxy) to filter content. These programs work by caching and filtering content before it is displayed in a user's browser. This provides an opportunity to remove not only ads but also content that may be offensive, inappropriate, or even malicious (Drive-by download). Popular proxy software which blocks content effectively include Netnanny, Privoxy, Squid, and some content-control software. The main advantage of the method is freedom from implementation limitations (browser, working techniques) and centralization of control (the proxy can be used by many users). Proxies are very good at filtering ads, but they have several limitations compared to browser-based solutions. For proxies, it is difficult to filter Transport Layer Security (SSL) (https://) traffic and full webpage context is not available to the filter. As well, proxies find it difficult to filter JavaScript-generated ad content. Hosts file and DNS manipulation Most operating systems, even those which are aware of the Domain Name System (DNS), still offer backward compatibility with a locally administered list of foreign hosts. This configuration, for historical reasons, is stored in a flat text file that by default contains very few hostnames and their associated IP addresses. Editing this hosts file is simple and effective because most DNS clients will read the local hosts file before querying a remote DNS server. Storing black-hole entries in the hosts file prevents the browser from accessing an ad server by manipulating the name resolution of the ad server to a local or nonexistent IP address (127.0.0.1 or 0.0.0.0 are typically used for IPv4 addresses). While simple to implement, these methods can be circumvented by advertisers, either by hard-coding, the IP address of the server that hosts the ads (this, in its turn, can be worked around by changing the local routing table by using for example iptables or other blocking firewalls), or by loading the advertisements from the same server that serves the main content; blocking name resolution of this server would also block the useful content of the site. Using a DNS sinkhole by manipulating the hosts file exploits the fact that most operating systems store a file with IP address, domain name pairs which is consulted by most browsers before using a DNS server to look up a domain name. By assigning the loopback address to each known ad server, the user directs traffic intended to reach each ad server to the local machine or to a virtual black hole of /dev/null or bit bucket. DNS filtering Advertising can be blocked by using a DNS server which is configured to block access to domains or hostnames which are known to serve ads by spoofing the address. Users can choose to use an already modified DNS server or set up a dedicated device running adequate software such as a Raspberry Pi running Pi-hole themselves. Manipulating DNS is a widely employed method to manipulate what the end-user sees from the Internet but can also be deployed locally for personal purposes. China runs its own root DNS and the EU has considered the same. Google has required their Google Public DNS be used for some applications on its Android devices. Accordingly, DNS addresses/domains used for advertising may be extremely vulnerable to a broad form of ad substitution whereby a domain that serves ads is entirely swapped out with one serving more local ads to some subset of users. This is especially likely in countries, notably Russia, India and China, where advertisers often refuse to pay for clicks or page views. DNS-level blocking of domains for non-commercial reasons is already common in China. Recursive Local VPN On Android, apps can run a local VPN connection with its own host filtering ability and DNS address without requiring root access. This approach allows adblocking app to download adblocking host files and use them to filter out ad networks throughout the device. AdGuard, Blokada, DNS66, and RethinkDNS are few of the popular apps which accomplish adblocking without root permission. The adblocking is only active when the local VPN is turned on, and it completely stops when the VPN connection is disconnected. The convenience makes it easy to access content blocked by anti-adblock scripts. This approach optimizes battery usage, reduces internet slowdown caused by using external DNS or VPN adblocking and needs overall less configuration. Hardware devices Devices such as AdTrap or PiHole use hardware to block Internet advertising. Based on reviews of AdTrap, this device uses a Linux Kernel running a version of PrivProxy to block ads from video streaming, music streaming, and any web browser, while PiHole act as local DNS to block advertisement servers, stopping connected devices from showing most ads. Another such solution is provided for network-level ad blocking for telcos by Israeli startup Shine. By external parties and internet providers Internet providers, especially mobile operators, frequently offer proxies designed to reduce network traffic. Even when not targeted specifically at ad filtering, these proxy-based arrangements will block many types of advertisements that are too large or bandwidth-consuming, or that are otherwise deemed unsuited for the specific internet connection or target device. Many internet operators block some form of advertisements while at the same time injecting their own ads promoting their services and specials. Economic consequences for online business Some content providers have argued that widespread ad blocking results in decreased revenue to a website sustained by advertisements and e-commerce-based businesses, where this blocking can be detected. Some have argued that since advertisers are ultimately paying for ads to increase their own revenues, eliminating ad blocking would only dilute the value per impression and drive down the price of advertising, arguing that like click fraud, impressions served to users who use ad blockers are of little to no value to advertisers. Consequently, they argue, eliminating ad blocking would not increase overall ad revenue to content providers in the long run. Business models Tools that help block ads have to work on different business models to stay in operation: Free and open source: Several tools work under a FOSS model, powered by community contributions and donations. Eg. uBlock Origin Whitelisting: Companies have resorted to maintaining a whitelist against a share of the ad revenue to allow "acceptable ads". This has faced criticism, such as AdBlock Plus Subscription/Upfront: Some companies in this field have started a subscription or upfront payment model for the tools. Eg. Wipr Freemium: Other companies offer some level of service for free while charge for additional features. Eg. AdGuard Response from publishers Countermeasures Some websites have taken countermeasures against ad blocking software, such as attempting to detect the presence of ad blockers and informing users of their views, or outright preventing users from accessing the content unless they disable the ad blocking software, whitelist the website, or buy an "ad-removal pass". There have been several arguments supporting and opposing the assertion that blocking ads is wrong. The back-and-forth elevation of technologies used for ad-blocking and countering ad-blocking have been equated to an "ad blocking war" or "arms race" between all parties. It has been suggested that in the European Union, the practice of websites scanning for ad blocking software may run afoul of the E-Privacy Directive. This claim was further validated by IAB Europe's guidelines released in June 2016 stating that there indeed may be a legal issue in ad blocker detection. While some anti-blocking stakeholders have tried to refute this it seems safe to assume that Publishers should follow the guidelines provided by the main Publisher lobby IAB. The joint effort announced by IAB Sweden prior to IAB Europe's guideline on the matter never materialized, and would have most likely been found against European anti-competition laws if it did. In August 2017, a vendor, Admiral, of such counter-measures issued a demand under section 1201 of the U.S. Digital Millennium Copyright Act, to demand the removal of a domain name associated with their service from an ad-blocking filter list. The vendor argued that the domain constituted a component of a technological protection measure designed to protect a copyrighted work, and thus made it a violation of anti-circumvention law to frustrate access to it. Alternatives As of 2015, advertisers and marketers look to involve their brands directly into the entertainment with native advertising and product placement (also known as brand integration or embedded marketing). An example of product placement would be for a soft drink manufacturer to pay a reality TV show producer to have the show's cast and host appear onscreen holding cans of the soft drink. Another common product placement is for an automotive manufacturer to give free cars to the producers of a TV show, in return for the show's producer depicting characters using these vehicles during the show. Some digital publications turned to their customers for help as a form of tip jar. For example, the Guardian is asking its readers for donations to help offset falling advertising revenue. According to the newspaper's editor-in-chief, Katharine Viner, the newspaper gets about the same amount of money from membership and donations as it does from advertising. The newspaper considered preventing readers from accessing its content if usage of ad-blocking software becomes widespread, but so far it keeps the content accessible for readers who employ ad-blockers. A new service called Scroll, launched in January 2020, worked with several leading website publishers to create a subscription model for ad-free browsing across all supported websites. Users would pay Scroll directly, and portions of the subscription fees are doled out to the websites based on proportional view count. See also Adware Adversarial information retrieval Content-control software Criticism of advertising Content filter Commercial skipping DNSBL Hosts file filtering Malvertising Privacy Proxomitron References Advertising-free media Anti-spam Online advertising Promotion and marketing communications
54524521
https://en.wikipedia.org/wiki/Strengthening%20State%20and%20Local%20Cyber%20Crime%20Fighting%20Act%20of%202017
Strengthening State and Local Cyber Crime Fighting Act of 2017
The Strengthening State and Local Cyber Crime Fighting Act of 2017 (H.R. 1616) is a bill introduced in the United States House of Representatives by U.S. Representative John Ratcliffe (R-Texas). The bill would amend the Homeland Security Act of 2002 to authorize the National Computer Forensics Institute, with the intent of providing local and state officials with resources to better handle cybercrime threats. Ratcliffe serves as the current chairman of the House Homeland Security Subcommittee on Cybersecurity and Infrastructure Protection. The bill was passed by the House with a roll call vote of 408-3 after forty minutes of debate. Between its introduction and approval, the bill was referred to the Committee on the Judiciary, the Committee on Homeland Security, the Subcommittee on Transportation and Protective Security, and the Subcommittee on Crime, Terrorism, Homeland Security, and Investigations. The bill has a total of 18 cosponsors, including 17 Republicans and one Democrat. Ratcliffe introduced the bill because he believes that local and state level law enforcement entities should be better equipped to handle emerging cyber threats in order to protect communities. He expressed concern that in today's world, traditional evidence of crimes, like DNA samples, might not be enough to solve cases, because criminals are more frequently breaking the law and leaving behind traces on the internet. In March 2017, Ratcliffe said, "Cyber elements add layers of complexity to the crimes our local law enforcement officers face every day ‒ and we've got to make sure they have access to the training they need to address this trend." As of July 2017, the Senate has not yet considered the bill, although Senators Chuck Grassley (R-Iowa), Dianne Feinstein (D-California), Richard Shelby (R-Alabama), Sheldon Whitehouse (D-Rhode Island), and Luther Strange (R-Alabama) introduced a companion bill. Senator Grassley, current Senate Judiciary Committee Chairman, supported the role of the National Computer Forensics Institute and the purpose of Ratcliffe's bill, saying the center gives officials the capacity to "dust for 'digital fingerprints' and utilize forensics to gather evidence and solve cases." Background As the internet has grown increasingly dominant in communications, business, and daily life, cybercrime has become a more dangerous threat to privacy and safety of both public and private entities. Americans recognize this issue, as they identified cyberattacks as the second most concerning global threat to the nation in a May 2016 survey conducted by Pew Research. The federal government has fallen victim to cyber attacks numerous times, especially in recent years. A 2016 report from the Government Accountability Office that included a survey of two dozen federal agencies found that cyberattacks exploded in frequency by 1,300 percent over the course of a decade, with 5,500 in 2006 to as many as 77,000 in 2015. The report listed eighteen of the agencies as operators of "high-impact systems", meaning that leaks could cause "catastrophic harm" to particular individuals, the government, or the general public. In May 2015, hackers accessed tax returns of over 300,000 people during an attack on the Internal Revenue Service, which led to the distribution of over $50 million in fraudulent refunds. Since 2007, federal agencies have dealt with at least a dozen attacks on their systems. Social Security information, home addresses, health records, and other personal data has been stolen from millions of people, including elected officials. Hackers from foreign countries have played a significant role in cyber attacks against the United States. The FBI, CIA, NSA and the Office of the Director of National Intelligence agreed that Russian hackers were behind an infiltration of the Democratic National Committee's data during the 2016 United States presidential election. NBC News obtained a document from the National Security Agency that detailed a five-year period of Chinese hacking of government information, which ended in 2014. Private groups and businesses have also been damaged by cybercrime. Cyberattacks cost American companies more than $15 million per year. The United States loses 500,000 jobs on an annual basis due to cyberattacks. More than 50 percent of all cyberattacks are launched against American companies, leading to a loss of $100 billion in the economy each year. A Citibank ATM server attack in 2008 lead to the loss of $2 million, and a Sony hack in 2014 lead to the loss of $15 million. In 2014 alone, 44% of all small businesses in the United States were attacked, which resulted in affected companies losing an average of $8,700 each. Ratcliffe and supporters of the bill aim to better equip all levels of the government to handle the increasing cyber threats that challenge both the government and American businesses and private citizens. U.S. Representative Dan Donovan (R-NY), who voted in favor of the bill, said, "It's critical that our law enforcement agencies have the information and training they need to respond to threats. This legislation will help increase coordination between the brave men and women who defend us on the local, state, and federal level and ensure that they have the tools needed to protect the homeland." The National Computer Forensics Institute has proven vital in this effort, having already worked with over 6,250 officials coming from all 50 states and even three United States territories. Representative Bob Goodlatte (R-VA), House Judiciary Committee chairman, said the NCFI would benefit from a clear congressional mandate. "Authorizing the existing National Computer Forensics Institute in federal law will cement its position as our nation's premier hi-tech cyber crime training facility." Major provisions The objective of the bill is to authorize the National Computer Forensics Institute so that the center can "disseminate homeland security information related to the investigation and prevention of cyber and electronic crime and related threats, and educate, train, and equip State, local, tribal, and territorial law enforcement officers, prosecutors, and judges." The bill aims to achieve this through amending the Homeland Security Act of 2002. While the National Computer Forensics Institute, based in Hoover, Alabama, has been open since 2008, the bill would ensure proper oversight and accountability, solidifying the center's position as the optimal site for training law enforcement to mitigate cyber threats. The bill establishes that the operations of the National Computer Forensics Institute are to: Educate law enforcement officers at all levels of government, including local, state, territorial, and tribal forces on threats related to cybercrime. Instruct law enforcement personnel, about proper investigation methods, device examination techniques, and potential prosecutorial and judicial challenges. Train law enforcement officers to carry out investigations and respond in a necessary manner to reported incidents of cybercrime. Facilitate the sharing of information between State, local, tribal and territorial officers and prosecutors. Provide materials like hardware, software, manuals, tools, and other items to law enforcement to conduct investigations. Expand the Electronic Crime Task Forces of the United States Secret Service through continuous training of officers. The NCFI would continue to be operated by the U.S. Secret Service, within the Department of Homeland Security. The network of Electronic Crime Task Forces would include stakeholders from academic institutions and the private sector under the expansion sanctioned by the authorization of the NCFI. Legislative history During the 114th Congress, an earlier version of the bill was passed in the U.S. House on November 30, 2015. On December 1, 2015, the bill was received in the Senate and referred to the Committee on the Judiciary. No further action was taken. On March 17, 2017, Ratcliffe introduced the bill in the House. After introduction: 3/17/17 - Referred to the Committee on the Judiciary and Committee on Homeland Security. 3/29/17 - Referred to the Subcommittee on Transportation and Protective Security by the Committee on Homeland Security. 3/31/17 - Referred to the Subcommittee on Crime, Terrorism, Homeland Security, and Investigations by the Committee on the Judiciary. 5/16/17 - The House passed the bill on motion to suspend the rules and pass the bill as amended with a vote of 403 Yeas and 8 Nays. 5/17/17 - The Senate received the bill, read it twice, and referred it to the Committee on the Judiciary. See also Cybercrime Homeland security References External links National Computer Forensics Institute Proposed legislation of the 115th United States Congress Cybercrime
23777555
https://en.wikipedia.org/wiki/Computer%20security%20conference
Computer security conference
A computer security conference is a convention for individuals involved in computer security. They generally serve as meeting places for system and network administrators, hackers, and computer security experts. Events Common activities at hacker conventions may include: Presentations from keynote speakers or panels. Common topics include social engineering, lockpicking, penetration testing, and hacking tools. Hands-on activities and competitions such as capture the flag (CTF). "Boot camps" offering training and certification in Information Technology. List of general computer security conferences General security conferences might be held by non-profit/not-for-profit/for-profit professional associations, individuals or informal group of individuals, or by security product vendor companies. 44Con An Infosec conference and training event that occurs annually in London, UK. ACM-CCS (Conferences on Computer and Communications Security), security conference held since 1993. ACSAC, Annual Computer Security Applications Conference - oldest information security conference held annually. ASIA or the Annual Symposium on Information Assurance that serves as the academic track for the New York State Cyber Security Conference, an annual information security conference held in Albany, NY usually for two days during June targeted at academic, government, and industry participants. Black Hat, a series of conferences held annually in different cities around the world. Black Hat USA, held in Las Vegas immediately before DEF CON, is one of the largest computer security events in the world. BlueHat Conference, a twice a year, invitation-only Microsoft security conference aimed at bringing Microsoft security professionals and external security researchers together. Brucon, yearly conference, previously held in Brussels, since 2012 it is held in Ghent. Lasting 2 days, and preceded by a training. CanSecWest, in Vancouver is held at the end of March and hosts the Pwn2Own hacking contest. CSS - International Conference on Cryptography and Security System in Poland. CIA Conference - A series of international conference on cyber security which aims to bridge the gap between the corporate need of cyber security to the fancy world of hacking. DeepSec, in Vienna covers many security aspects of computing and electronic communications as well as security management and social aspects. DeepSec is visited by a broad international audience, academics, researchers, vendors, finance, public administration etc. (2 days trainings, 2 days conference). Presentations are published on Vimeo and YouTube. DEF CON-is one of the world's largest and most notable hacker conventions, held annually in Las Vegas, Nevada. Department of Defense Cyber Crime Conference, an annual conference that focuses on the computer security needs of the United States federal government, military, and defense contractors. FSec - Croatian annual security conference held at the Faculty of organization and informatics in Varaždin. GreHack.fr an annual conference held in Grenoble. Speakers from academia, industry. Both offensive and defensive security. Hack.lu, an annual conference held in Luxembourg Hacker Halted, Presented by EC-Council, the objective of the global series of Hacker Halted conferences is to raise international awareness towards increased education and ethics in IT Security. HackinParis, Is an annual event organized by SYSDREAM. The program includes IT security, industrial espionage, penetration testing, physical security, social engineering, forensics, malware analysis techniques and countermeasures. HACK-IT-N is a cybersecurity conference that deals with current issues and offensive and defensive techniques of the moment: hacking, cyber-espionage, cyber-surveillance etc. It's held in Bordeaux at the end of each year and organized by TEHTRIS and the ENSEIRB-MATMECA engineering school. Hackito Ergo Sum, Security conference pertaining to research topics, with attendees and speakers from both the industry, the offensive side and the academic circles, held in Paris every April. HITBSecConf / Hack In The Box, deep-knowledge security conference held in Malaysia and the Netherlands. ICISSP International Conference on Information Systems Security and Privacy, IEEE Symposium on Security and Privacy, top-tier privacy & security conference. IIA/ISACA Chicago IT Hacking and Cybersecurity Conference, a regional, two-day event that brings together executives and professionals in audit, IT, and risk management functions to hear current trends and first-hand knowledge from subject matter experts. INFOWARCON Beyond Information Warfare - Offensive Cyber Weapons and Technologies Training Congress. IP EXPO Europe, held annually at London's ExCeL Centre IP EXPO Nordic, held annually at Stockholm's Waterfront Congress Centre IWSECC International Workshop on Security Engineering for Cloud Computing in conjunction with the 14th International Conference on Availability, Reliability and Security (ARES Conference » Vienna, Austria) LeetCon, IT-Security-Convention in Hannover (Germany), frequently October or November every Year. Talks about IT-Security, IoT, Industry 4.0 and more. NDSS (Network & Distributed System Security Symposium), annual security conference from Internet Society. NSPW (New Security Paradigms Workshop), a workshop with new ideas in security. nullcon Security Conference is an annual security event held in Goa, India. Open Web Application Security Project (OWASP), Focuses on web application security. REcon REcon is a computer security conference with a focus on reverse engineering and advanced exploitation techniques. RSA Security Conference, Cryptography and information security-related conference held annually in the San Francisco Bay Area. RuhrSec, annual non-profit security conference at the Ruhr University Bochum. S4:SCADA Security Scientific Symposium, Security conference pertaining to SCADA held annually by Digital Bond, usually in Miami. SecurIT 2012, International Conference on Security of Internet of Things held in mid of August at Amrita Vishwa Vidyapeetham. SecureWorld Expo, a series of IT Security conferences offering education, CPE training opportunities, and networking of security leaders, experts, senior executives, and policy makers who shape the face of security. SIN ACM, or the International Conference on Security of Information and Networks. SOURCE Conference, SOURCE is a computer security conference in Boston, Dublin and Seattle that offers education in both the business and technical aspects of the security industry. , Annual French Security Symposium held in Rennes. Swiss Cyber Storm, International IT Security Conference held in October at KKL Lucerne. TROOPERS IT Security Conference, Annual international IT Security event with workshops held in Heidelberg, Germany. USENIX Security, annual security conference associated with USENIX. Virus Bulletin conference, annual security conference held late September or early October since 1989. 0SecCon Zero Security Conference held in Kerala. Positive Hack Days, Annual international IT Security event with workshops held in Moscow, Russia. IEEE WIFS Workshop on Information Forensics and Security conference hosted by IEEE Signal Processing Society. Hacker conferences A hacker conference, also known as a hacker con, is a convention for hackers. These serve as meeting places for phreakers, hackers, and security professionals. The actual events, time-spans, and details of various themes of these conventions not only depends on the specific convention attended but also its perceived reputation. Typically the actual details of any given convention are couched in mild secrecy due to the legality of certain panels, as well as the willingness of attendees to explain themselves to law enforcement and less computer-savvy individuals (see hacker definition controversy). Common topics include wardriving, lockpicking, corporate and network security, personal rights and freedoms, new technologies, as well as general 'geek' motifs. Some may also have contests and general collaborative events such as hackathons. One facet of Hacker conferences that tends to differentiate many "HackerCons" from general computer security conferences, is the evolution of the conference badge. While many conferences use a fairly standard paper/plastic badge to identify attendees by name, many hacker con's have evolved to use more non-traditional badges, such as electronic PCB's with LED's, LCD screens, and can include wifi and advanced, often hidden functionality, to include games, that do not identify the individual, sometimes promoting a group identity or regional/local group affiliation. This has spurred the creation of so-called "badgelife" where different individuals and organizations (such as local hacker groups) may design, create, and sell or otherwise distribute a separate badge attendees wear in addition to their conference badge. These badges may be purchased, but some may have to be earned by completing a challenge at the conference. One of the rarest and most desirable badges is the DefCon Black Badge. While the use of the badges are traditionally associated with DefCon, their use has spread to other conferences. List of hacker conferences AthCon, the largest hacker conference in South Eastern Europe taking place annually in Athens, Greece. BSides, community conference that initially started in the US is now global. Blackhoodie, is a series of free, women-only reverse engineering conference offering workshops and talks. They are held in different cities around the world on a regular basis to encourage women interested in the cyber security field. CactusCon, the largest hacker and cybersecurity conference in Arizona, held annually since 2012. CarolinaCon, an annual information security and rights conference held every spring in North Carolina's Research Triangle. Chaos Communication Congress, the oldest and Europe's largest hacker conference, held by Chaos Computer Club. CIA Conference - A series of international conference on cyber security which aims to bridge the gap between the corporate need of cyber security to the fancy world of hacking. CircleCityCon is a security and technology conference held annually in June in Indianapolis. Cop CON, is a unique Cyber Policing Conference held across India every year by Indian Cyber Army. CypherCon, a Milwaukee based hacker conference held late winter each year. DEF CON, in Las Vegas, Nevada, is the biggest hacker convention in the United States held during summer (June–August). DerbyCon, an annual hacker conference based in Louisville, KY (2011-2019). Diana Initiative, an annual hacker conference to promote Women in cybersecurity technologies by The Diana Initiative non profit org located in Las Vegas, NV. Digital Overdose Con, a Europe based hacker and tech conference focused on rookies held in the spring each year. Ekoparty, a hacker convention in Argentina and one of the most important in South America, held annually around September in Buenos Aires. GrrCon, an annual hacker conference hosted each September in Grand Rapids, Michigan. GroundZero Summit, an annual hacker conference hosted each November in Hotel Ashoka, New Delhi. Hack in the Box, an annual hacker conference. Hackers Day is one of the hacking conferences in India held frequently in January–February every year. The organizers also used to organize DEFCON Lucknow, which was dissolved by the DEFCON after many controversies arose in 2016. Hackers 2 Hackers Conference (H2HC) is the oldest security research (hacking) conference in Latin America and one of the oldest ones still active in the world. Hackers Idol, is a unique Hackers & Cyber Enthusiast Talent Hunt held across India every year by Indian Cyber Army. Hackers Next Door, in New York City is held by Tech Learning Collective during the Winter (November/December). Hackers on Planet Earth (HOPE), in New York City is held by 2600: The Hacker Quarterly in mid-summer (July/August) every other year. Hackfest.ca, in Quebec, Canada, bilingual conferences and hacking games. Hackers Teaching Hackers, a hacker conference in Columbus, Ohio, featuring talks, hands-on villages, a CTF, mentorship, and more. HackMiami Conference, a hacker conference in Miami, Florida, organized by the HackMiami hackerspace. Hacktivity, in Budapest, Hungary, run every September, the largest hacker conference in the Central and Eastern Europe. Hardwear.io - Hardware Security Conference is an annual Hardware Security event held in The Hague, Netherlands. It is being organized every year to answer the emerging threats and attacks on hardware. INFILTRATE, hosted by Immunity, Inc, is a deep technical security conference that focuses on offensive technical issues. The conference has been held annually in Miami Beach, Florida since 2011. Kernelcon, is an Omaha, Nebraska hacker convention held annually in April, beginning in 2019. Kiwicon, is a Wellington, New Zealand hacker convention. LayerOne, held every spring in Los Angeles, California. LeHack, FRENCH HACKERS CONFERENCE hosted in Paris Malcon, the world's first International Malware Conference, hosted in India. No Hat is an international computer security conference held annually in Bergamo (IT), and considered the largest technical security event in Italy. NorthSec, in Montreal, Quebec, is an applied security event held yearly mid-May, featuring two days of conference followed by a 48h on-site CTF. Notacon, in Cleveland, Ohio, is an art and technology conference held frequently in mid-April. Nuit Du Hack, in Paris, France, is the oldest and largest hacker conference held frequently in late-June. PhreakNIC, in Nashville, Tennessee, is held by Nashville 2600. around October. Pumpcon, A yearly hackercon in a bar in Philadelphia, PA. Quahogcon, In Providence, Rhode Island is held at the end of April. Roadsec, One of the biggest hacking conferences in Latin America, held in Brazil, with smaller city editions throughout the whole year, and a big ending edition held in São Paulo in November (February - November). ROOTCON, Premier Annual Hacker Conference in the Philippines held annually during summer (September - October). RootedCON, in Madrid, Spain, is one of the biggest hacker conference in Europe. This convention started in 2010. (March) ShellCon, in Los Angeles CA is an information security conference that is held annually in the beautiful beach cities of Los Angeles. ShmooCon, in Washington, DC is an annual east coast hacker conference that is held in Washington, DC. SkyDogCon, A technology conference in Nashville, TN for the individual with the Renaissance Mind. SkyDogCon exists to facilitate learning, information sharing, and mingling with like-minded people in a relaxed atmosphere. Summercon, one of the oldest hacker conventions, held during Summer (frequently in June). It helped set a precedent for more modern "cons" such as H.O.P.E. and DEF CON. The Hackers Conference, is India's biggest Cyber Security Conference held in New Delhi, India every Year. THOTCON, a Chicago based hacker conference held in the Spring each year. ToorCon, San Diego hacker convention that emerged from the 2600 user group frequently in late September. WildWestHackinFest, a conference focused on training and development held in Deadwood, South Dakota in October. Explore the Black Hills and learn how to hack all the things, including the IoT. ZeroAccess, is a unique International Cyber Security Conference held in India, India in the Month of April by Indian Cyber Army Sthack is an annual security conference in Bordeaux, followed by the traditional Capture The Flag. Non-annual hacker conventions Stichting HAL2001, a not-for-profit organization, holds a quadrennial Dutch hacker convention. They have, so far, held HAL2001 (Hackers at Large) and What the Hack (2005, originally called HEX (referring to the 16th anniversary of the event, as well as an acronym for Hacking Extreme)). The Dutch conferences held prior Stichting HAL2001's existence were Galactic Hacker Party (1989), Hacking at the End of the Universe (1993) and Hacking In Progress (1997), notable for being held simultaneously and in conjunction with Beyond HOPE. The not-for-profit organization 'Stichting Hxx' was organizing the 2009 incarnation of this con; HAR (Hacking at Random). During the summer of 2013 the most recent installment of the series, called OHM (Observe Hack Make), was held. References External links Hacker Convention Comparison based on internet connectivity and attendees InfoSec conferences List of french non-commercial IT sycurity event in 2019. Hacker conventions
9812036
https://en.wikipedia.org/wiki/FAIR%20USE%20Act
FAIR USE Act
The "Freedom and Innovation Revitalizing United States Entrepreneurship Act of 2007" (FAIR USE Act) was a proposed United States copyright law that would have amended Title 17 of the U.S. Code, including portions of the Digital Millennium Copyright Act (DMCA) to "promote innovation, to encourage the introduction of new technology, to enhance library preservation efforts, and to protect the fair use rights of consumers, and for other purposes." The bill would prevent courts from holding companies financially liable for copyright infringement stemming from the use of their hardware or software, and proposes six permanent circumvention exemptions to the DMCA. The bill was introduced February 27, 2007 in the 110th Congress by Representative Rick Boucher (D-VA). On March 19, 2007, the bill was referred to the House Subcommittee on Courts, the Internet, and Intellectual Property. The bill was not reintroduced. Boucher emphasized that the bill would not make circumvention an act of fair use, but would instead redefine which acts qualify as permissible circumvention, stating that “The Digital Millennium Copyright Act dramatically tilted the copyright balance toward complete copyright protection at the expense of the Fair Use rights of the users of copyrighted material. The reintroduced legislation will assure that consumers who purchase digital media can enjoy a broad range of uses of the media for their own convenience in a way which does not infringe the copyright of the work.” History The FAIR USE Act is Boucher’s third attempt at reforming provisions within the DMCA, the previous two being the Digital Media Consumers' Rights Acts (DMCRA) of 2003 and 2005. Previously, Boucher co-sponsored the “Benefit Authors without Limiting Advancement or Net Consumer Expectations,” or “BALANCE Act,” which sought to amend the DMCA to account for noninfringing circumvention. The Digital Media Consumers' Rights Acts stressed the necessity of adequate labeling on media bearing digital rights management (DRM) and similar protections to prevent consumer confusion. Both amended the DMCA to include exceptions for acts of circumvention that furthered consumers’ exercise of fair use rights. The DMCRA of 2003 included a section of fair use amendments, including amendments to the exemptions described by Section 1201(c) of Title 17. The bill exempted research into “technological measures” from infringement and enabled consumers to circumvent DRM, and qualified that using services for noninfringing uses would not be a violation. The bill also explicitly stated that manufacturing and distributing hardware or software capable of noninfringing uses would likewise not be a violation. The revised DMCRA of 2005 included a similar section of "fair use amendments", but did not make mention regarding users of noninfringing circumvention services. The Motion Picture Association of America (MPAA) and the Recording Industry Association of America (RIAA) criticized both incarnations of the bill, arguing that the language was too permissive and would “legalize hacking and piracy.” Provisions The FAIR USE Act consists of two main provisions: redefining copyright infringement and amending copyright circumvention exemptions. Section 2: Copyright infringement Section 2 would amend Section 504(c)(2) of Title 17, and would prevent courts from levying statutory damages in cases of secondary infringement. Section 2 would also amend Section 501 by adding that "no person shall be liable for copyright infringement based on the design, manufacture, or distribution of a hardware device or of a component of the device if the device is capable of substantial, commercially significant noninfringing use." In effect, this would reverse the Supreme Court’s decision in MGM Studios, Inc. v. Grokster, Ltd. (2005), which held that "one who distributes a device with the object of promoting its use to infringe copyright ... is liable for the resulting acts of infringement," which has gradually come to be interpreted to mean that any distribution of an object capable of infringement is liable for any resulting infringement. With the Grokster decision null, the prevailing standard regarding secondary liability would return to the Court's decision in Sony v. Universal (also known as the Betamax case). Section 2 echoes the Court's language in Sony, stating specifically that one may not be held liable for copyright infringement "based on the design, manufacture, or distribution of a hardware device or of a component of the device if the device is capable of substantial, commercially significant noninfringing use." Section 3: DMCA amendments Section 3 first codifies the set of circumvention exemptions granted by the Librarian of Congress as part of the 2006 DMCA rule making process. Section 3 also amends the DMCA to add exceptions for six types of circumvention. Circumvention by libraries and archives, to skip objectionable content, to transmit over a personal network, to gain access to public domain works, for public interest work and research, and for preservation are added as a new set of exceptions. (I) Libraries and archives Section (I) allows libraries and archives to circumvent copyright for the purposes of compiling audiovisual works that are in a library's collection for educational classroom use by an instructor. Educational compilations for college film and media courses were exempt by the Librarian of Congress under Section 1201 of the DMCA, but that exemption was set to expire in 2009. Section(I) not only made the exemption permanent, it expanded the exemption to apply to compilations for coursework at all grade levels, in any subject area. (II) Objectionable content Section (II) allows specifically for circumvention via hardware or software that skips objectionable content. Circumvention for the purposes of avoiding objectionable content became an issue in 2006, when a Denver judge ruled that the edited versions of films sold by companies such as CleanFlicks and CleanFilms were not considered fair use. These companies, along with a handful of others, removed objectionable content (such as nudity and profanity) from DVDs and sold the edited versions to consumers. In its decision, the court did not address the legality of companies who offered software or hardware that would “read” unaltered media and skip objectionable content. Section (II) allows the sale of hardware, such as modified DVD players sold by CleanPlay, and software, like downloadable plug-ins, that would skip such content. (III) Personal network Section (III) allows circumvention for the purpose of storing or transmitting media over a personal network, but explicitly prevents the uploading of media “to the Internet for mass, indiscriminate redistribution." (IV) Public domain works Section (IV) allows for circumvention that enables access to a public domain work, or a compilation of works that are primarily in the public domain. Arguably, Section (IV) would expressly allow initiatives such as Google Books, which was originally pioneered in 2004 as a database to increase the availability of, and readers' access to, public domain works. (V) Public interest work and research Section (V) is similar to a broader version of the third prong of fair use. It allows circumvention that is carried out to gain access to a work of substantial public interest solely for the purposes of "criticism, comment, news reporting, scholarship, or research." The language of Section (V) is ambiguous, which led some critics to worry that the language was too broad, potentially enabling students to circumvent copyright to access books, films, and music for coursework, or allowing professors to create course packs without obtaining permission from publishers. (VI) Circumvention for preservation Section (VI) allows circumvention for purposes of preservation by a library or archives, with respect to works in its collection. Since the DCMA was passed, librarians across the country protested the limited circumvention rights they were afforded. The Association of Research Libraries addressed the problem of preservation head-on, arguing that “preservation is one of a library’s most critical functions... the DMCA is interfering with our ability to preserve these works.” The provisions of Section (VII) would effectively eliminate librarians' problems with preservation of works in a library's collection. Criticism The FAIR USE Act was subject to criticism of both proponents and opponents of DMCA reform. Proponents of DMCA reform critiqued the bill for being "wishy-washy". Despite the included amendments, several of which arose in response to public protest, the Act did not allow for circumvention for ripping personal copies for cross-media consumption (i.e. ripping a DVD for use on a video phone or laptop), which some considered the biggest problem with the DMCA. The bill was also criticized for not maintaining the more strongly worded exemptions enumerated in previous incarnations of DMCA reform legislation, in particular, those regarding the makers and distributors of circumvention technology, which meant that “a film studies professor would be permitted to use software such as Handbrake... However, developing or distributing Handbrake in the United States would still be a crime.” Opponents of the bill focused on problems with the scope and breadth of its language. As with previous Boucher-sponsored bills, the RIAA argued that the bill would “repeal the DMCA and legalize hacking." See also Digital Millennium Copyright Act Digital Media Consumers' Rights Act INDUCE Act Public Domain Enhancement Act PRO-IP Act References External links FAIR USE Act The US Library of Congress page on this bill: The OpenCongress page on this bill: The GovTrack progress on this bill: United States Code Title 17, Chapter 12, § 1201: Title 17, Chapter 12, § 504: Proposed legislation of the 110th United States Congress United States proposed federal intellectual property legislation United States federal copyright legislation Fair use
41585692
https://en.wikipedia.org/wiki/Code.org
Code.org
Code.org is a non-profit organization and eponymous website led by Hadi and Ali Partovi that aims to encourage people, particularly school students in the United States, to learn computer science. The website includes free coding lessons and the initiative also targets schools in an attempt to encourage them to include more computer science classes in the curriculum. On December 9, 2013, they launched the Hour of Code nationwide to promote computer science during Computer Science Education Week through December 15, 2013. History Code.org was launched in January 2013 by Iranian-American brothers Hadi Partovi and Ali Partovi, as a non-profit focused on making computer programming more accessible. The initial focus was on creating a database of all computer science classrooms in the United States. At the time, Hadi Partovi stated that about ninety percent of US schools do not teach programming, despite it now being a "foundational field". The idea for the organization came from Hadi, who states that he thought of it on the day of Steve Jobs' death in 2011 while mulling over his own potential legacy. After this, Hadi worked full-time to bring the organization to fruition. In late February 2013, a month after launch, they released a video featuring Mark Zuckerberg, Bill Gates, Jack Dorsey, and other programmers and entrepreneurs on the importance of learning how to code. Two weeks after the launch, TechCrunch reported that the video had gone viral and received a lot of positive attention. Partovi raised about $10 million for Code.org from various tech companies and tech company founders. In 2014, Code.org posted a one-hour tutorial to build and customize a Flappy Bird video game using the site's block visual programming language. Code.org has also created coding programs revolving around characters from the Disney film Frozen, in addition to Angry Birds, and Plants vs. Zombies. In December 2014, Code.org held a successful Indiegogo crowdfunding campaign that raised over $5 million to help educate school children. Overall, about 100 million students have been reached by Code.org efforts. On November 16, 2015, Microsoft announced a partnership with Code.org to launch Minecraft as a tutorial to teach kids how to code. In February 2016, Code.org won the award for "Biggest Social Impact" at the Annual Crunchies Awards. By September 2016, Code.org founder Hadi Partovi was recognized as a Top 30 Education Leader by the Center for Digital Education. In December 2016, Code.org rang the opening bell of the NASDAQ. In March 2017 expanding its operation, it has started its working in the region of Rajasthan, India. Hour of Code During Computer Science Education Week from December 9 to December 15, 2013, Code.org launched the "Hour of Code Challenge" on its website to teach computer science to school students, enticing them to complete short programming tutorials. The Hour of Code involved getting people to write short snippets of code to achieve pre-specified goals using Blockly, a visual programming language of a similar flavor as Logo. The initiative had been announced about two months in advance and at the time of launch, the initiative was supported by former United States President Barack Obama as well as executives from technology companies such as Microsoft and Apple Inc. It was also supported by educational online learning platforms such as Khan Academy. About 20 million people participated and over 600 million lines of code had been written. The Hour of Code also offered participation gifts to some of the schools involved, such as a set of fifty laptops or a conference call with one a tech "luminary" like Gates or Dorsey. The crowdfunding effort for Hour of Code received the highest amount of funding ever raised on Indiegogo. By October 2014, about forty million students had taken the Hour of Code class, and a second Hour of Code was held in December 2014. That year, locations for Hour of Code lessons included Apple retail stores. In December 2016, Canadian Prime Minister Justin Trudeau helped launch the international Hour of Code campaign for 2016. And in December 2017, Code.org announced that Hour of Code had reached over 500 million hours served. Curriculum efforts The first step in the organization's efforts as regards the curriculum of schools was to work with US school districts to add computer programming as a class. Most US schools did not have a course code for computer sciences, in order for schools to be able to offer coding as a class. After this, the next step was to create free online teaching and learning materials for schools to use if instituting computer science classes. By 2014, Code.org had launched computer courses in thirty US school districts to reach about 5% of all the students in US public schools (about two million students), and by 2015, Code.org had trained about 15,000 teachers to teach computer sciences, able to reach about 600,000 new students previously unable to learn computer coding, with large percentages of those being either female or minorities. To date, Code.org has prepared over 72,000 educators to teach computer science. That year Code.org was partnered with about seventy of the largest US school districts (including each of the seven largest), representing several million students. The company also partnered with other computer class businesses and private entities to provide additional computer learning materials and opportunities. As of 2015, six million students had been enrolled in Code.org curriculum classes. That year, Code.org partnered with College Board, in order to develop a slate of advanced placement computer classes. The main platform used in Code.org instruction is Code Studio which according to TechCrunch, "teaches the underlying concepts in programming through the manipulation of blocks of logic that when stacked together in a particular order, move a character around a scene or draw a shape." In 2018, Code.org celebrated record participation by girls and underrepresented minorities in AP computer science classrooms, driven in large part by students in the Code.org CS Principles classrooms. Legislative efforts One of the main lobbying efforts of Code.org in state legislatures is to ensure that computer classes are not registered as foreign language classes, but as science classes, in order to ensure that more than coding is taught in computer courses. Code.org also focuses specifically on female and minority students, as the organization believes these are the students most at risk of not receiving computer science education before high school or college. Goals According to its website, Code.org has the following goals: Bringing Computer Science classes to every K–12 school in the United States, especially in urban and rural neighborhoods. Demonstrating the successful use of online curriculum in public school classrooms Changing policies in all 50 states to categorize C.S. as part of the math/science "core" curriculum Harnessing the collective power of the tech community to celebrate and grow C.S. education worldwide Increasing the representation of women and students of color in the field of Computer Science. Increasing their global outreach by making their courses available in more than 45 languages that are used in over 180 countries. Reception Writing for San Jose Mercury News, Mike Cassidy praised Code.org and the Hour of Code, writing: "A publicity stunt is what we need." John Dvorak was critical of the Hour of Code in an article for PC Magazine. Dvorak wrote: "I see it as a ploy to sell more computers in schools." See also Codecademy CodeCombat CodeHS Coursera edX Khan Academy Livecoding.tv LinkedIn Learning Scratch (programming language) Udacity Treehouse (company) Tynker References External links Programming games Non-profit organizations based in Seattle American educational websites Education companies established in 2013 2013 establishments in Washington (state) 501(c)(3) organizations Hadi Partovi
26050706
https://en.wikipedia.org/wiki/Voreen
Voreen
Voreen (volume rendering engine) is an open-source volume visualization library and development platform. Through the use of GPU-based volume rendering techniques it allows high frame rates on standard graphics hardware to support interactive volume exploration. History Voreen was initiated at the Department of Computer Science at the University of Münster, Germany in 2004 and was first released on 11 April 2008 under the GNU general public license (GPL). Voreen is written in C++ utilizing the Qt framework and using the OpenGL rendering acceleration API, and is able to achieve high interactive frame rates on consumer graphics hardware. It is platform independent and compiles on Windows and Linux. The source code and documentation, and also pre-compiled binaries for Windows and Linux, are available from its website. Although it is intended and mostly used for medical applications, any other kind of volume data can be handled, e.g., microscopy, flow data or other simulations. Concepts The visualization environment VoreenVE based on that engine is designed for authoring and performing interactive visualizations of volumetric data. Different visualizations can be assembled in form of so-called networks via rapid prototyping, with each network consisting of several processors. Processors perform more or less specialized tasks for the entire rendering process, ranging from supplying data over raycasting, geometry creation and rendering to image processing. Within the limits of their respective purposes, the processors can be combined freely with each other, and thereby granting a great amount of flexibility and providing a uniform way of handling volume rendering. Authors who need to implement a certain rendering technique can confine their work basically on the development of new processors, whereas users who only want to access a certain visualization simply need to employ the appropriate processors or networks and do not need to care about technical details. Features Visualization Direct volume rendering (DVR), isosurface rendering, maximum intensity projection (MIP) Support of different illumination models (Phong reflection model, toon shading, ambient occlusion) Large (out-of-core) data visualization (using an OpenCL octree raycaster) Streamline-based vector field visualization Multimodal volume rendering Geometry rendering with support for order-independent transparency Flexible combination of image processing operators (depth darkening, glow, chromadepth, edge detection) Visualization of time-varying as well as segmented 3D datasets Support for 1D and 2D transfer functions/CLUTs Configurable views for building more complex applications (triple view/quad view/tabbed view/splitter) Plotting Volume Ensemble visualization (similarity plot, ensemble mean/variance, parallel coordinates) Volume Processing Isosurface extraction Efficient basic 3D-image processing for very large (out-of-core) volumes Very large volume analysis (connected components, vessel network analysis) Interactive volume segmentation (random walker-based, vesselness filtering, basic thresholding) Interactive volume registration (manual or landmark-based) Vector field volume processing (Jacobian matrix, Delta/Q/Lambda2 vortex criterion, coreline extraction) Out-of-core processing of spatio-temporal multi-field ensemble datasets (ensemble analysis) Interaction Configurable application mode for improving usability for domain experts Axis aligned and arbitrarily aligned clipping planes Editors for 1D and 2D transfer functions Inspection of intermediate results Distance measurements Data I/O Support for several volume file formats (e.g. DICOM, TIFF stacks, HDF5, RAW, NetCDF, VTI, NIfTI-1) High-resolution screenshot and camera animation generation with anti-aliasing FFmpeg-based video export Python scripting for offline image processing and visualization Geometry in/export (e.g. for Additive Manufacturing) See also Volume rendering Volume ray casting References External links Voreen website Free data visualization software Free 3D graphics software Free software programmed in C++ 3D imaging Free DICOM software Free bioimaging software Science software that uses Qt Science software for Linux
4765142
https://en.wikipedia.org/wiki/Lattice-based%20access%20control
Lattice-based access control
In computer security, lattice-based access control (LBAC) is a complex access control model based on the interaction between any combination of objects (such as resources, computers, and applications) and subjects (such as individuals, groups or organizations). In this type of label-based mandatory access control model, a lattice is used to define the levels of security that an object may have and that a subject may have access to. The subject is only allowed to access an object if the security level of the subject is greater than or equal to that of the object. Mathematically, the security level access may also be expressed in terms of the lattice (a partial order set) where each object and subject have a greatest lower bound (meet) and least upper bound (join) of access rights. For example, if two subjects A and B need access to an object, the security level is defined as the meet of the levels of A and B. In another example, if two objects X and Y are combined, they form another object Z, which is assigned the security level formed by the join of the levels of X and Y. LBAC is also known as a label-based access control (or rule-based access control) restriction as opposed to role-based access control (RBAC). Lattice based access control models were first formally defined by Denning (1976); see also Sandhu (1993). See also References Computer security models Lattice theory Access control
35259975
https://en.wikipedia.org/wiki/RedSleeve
RedSleeve
RedSleeve is a free operating system distribution based on the Linux kernel. It is derived from the Red Hat Enterprise Linux (RHEL) distribution, ported to the ARM architecture. RedSleeve is derived from the free and open-source software made available by Red Hat, Inc., but is not produced, maintained or supported by Red Hat. Specifically, this product is built from the source code for Red Hat Enterprise Linux versions, under the terms and conditions of Red Hat Enterprise Linux's EULA and the GNU General Public License. The name RedSleeve was chosen because it makes it intuitively obvious what upstream distribution it is derived from, while at the same time implying that it is for the ARM platform (we have sleeves covering our ARMs). RedSleeve is different from other Red Hat Enterprise Linux derivatives such as CentOS and Scientific Linux in that it is also a port to a new platform (ARM) that is not supported by the upstream distribution. History The first official Alpha release was made available on February 12, 2012. Releases Versioning scheme RedSleeve aims to maintain equal versioning with the upstream distribution, both in terms of distribution release numbering and the individual package release numbering. The only exception is with the packages that had to be modified from the upstream release. This is only done to either remove upstream branding as required by the upstream distributions terms and conditions, or to apply additional patches required to make the package build and work on the ARM architecture which at the time of writing the upstream distribution does not support. In such cases, .0 is appended to the package version after the distribution tag. Media coverage RedSleeve Linux 6 was covered on The Register on May 29, 2012. References ARM architecture ARM Linux distributions
949269
https://en.wikipedia.org/wiki/Soccer%20Kid
Soccer Kid
Soccer Kid is a 1993 side-scrolling platform video game originally developed and published by Krisalis Software in Europe for the Amiga. In the game, players assume the role of the titular main protagonist who travels across several countries around the world in order to repair the World Cup by retrieving pieces that were scattered by the alien pirate Scab, the main antagonist who failed to steal and add it to his trophy collection in a robbery attempt. Its gameplay mainly consists of platforming and exploration elements, with a main single-button or two-button configuration, depending on the controls setup. Conceived by Dean Lester, Soccer Kid was created by most of the same team who previously worked on the popular Manchester United franchise at Krisalis Software and uses the same game engine as with Arabian Nights, another title made by the developer. After making multiple association football titles, the team experimented with creating a project that fused both football and platform game elements in 1992, developing a physics engine dedicated to the soccer ball that proved to be successful internally. Initially released for the Amiga platform, the title was later ported to other home computers and consoles including the 3DO Interactive Multiplayer, Amiga CD32, Atari Jaguar, Game Boy Advance, MS-DOS, PlayStation and Super Nintendo Entertainment System, each one featuring several changes and additions compared to the original version. Conversions for multiple systems were also in development but never released. Since its original release on the Amiga, Soccer Kid garnered positive reception from critics who praised multiple aspects such as the presentation, visuals, original gameplay concept and replay value, while some felt divided in regards to the difficulty and sound design, with others criticized some of the repetitive set pieces later in the game. The Super Nintendo version received a more mixed reception from reviewers who felt very divided with the graphics, sound design and gameplay. The DOS conversion, although mixed, got a critical response similar to the original Amiga version for its visuals, sound and gameplay. The CD32 port was very well received by critics and the 3DO port got mostly positive reviews. The Jaguar port was received with mixed opinions from reviewers who criticized and felt that it did not improve the graphics and audio from the 16-bit versions, while the Game Boy Advance release was received with a more warm reception. Gameplay Soccer Kid is a platform game where players take control of the titular character travelling across the world in order to retrieve and restore the world cup from the clutches of the alien pirate Scab, who plans to add it to his ever-growing collection. By pressing different buttons at the title screen, the player can change the colours of their character's clothes in order to represent their favorite team. Each country has three levels that the player must navigate through. The player character can perform various types of soccer moves such as runs, shots, bicycle kicks, headers and other sorts of moves to either advance in the level, get to hard-to-reach areas or eliminate enemies by using his soccer ball as the main tool. The player starts out with two hearts but by opening random chests scattered throughout the game, they can get more hearts. At the end of each third level, the player must fight a boss, based on stereotypical people associated with their respective country. However, the player must always explore each level to find soccer cards, which are crucial to getting the cup at the end of the game. Once a country is completed and all the cards in that country are collected, Soccer Kid is transported to a bonus level where he must collect all the food against limited time in order for a piece of the cup to be obtainable. If all heart pieces are lost, the player character is respawned on a determined checkpoint after losing a life and once all lives are lost, the game is over, though there is an option of continuing. Depending on the version being played, progress is saved differently. Development Soccer Kid was created by most of the same personnel who worked on the Manchester United series including programmer Peter Harrap, who co-wrote the title alongside Nigel Little, artist Neil Adamson and composer Matt Furniss, among other team members that shared multiple roles during its development process, while the overall concept was conceived by Dean Lester, in addition to also sharing the same game engine as with Arabian Nights, another platform game by Krisalis. Development of the project began in February 1992, when Krisalis co-founder Tony Kavanagh announced their plans for it early in the year and the team wanted to experiment in making a title that mixed football and platforming elements after developing several association football titles. Nigel stated that in order to assure such project would be playable, the team wrote a primitive physics engine early in its development and implemented real-life rules for the ball to react realistically, which would eventually prove to be feasible after multiple testing phases. Soccer Kid was previewed across several video game magazines early in development under the title Football Kid, featuring different visuals compared with the final release, which were improved from its original incarnation, while some publications compared it with Sonic the Hedgehog due to its gameplay structure and console-style presentation. The game runs at 50 frames per second, while sprites were made as 64x64 pixel objects. It was originally intended to be released in November 1992 but due to the interest in porting the title to home consoles, the title was moved to next year and faced constant delays prior to its eventual launch. The name of the project was also changed in order to appeal to consumers when porting across other platforms. Release Soccer Kid was first showcased to attendees of Spring ECTS '93 and it was first released for the Amiga in Europe in June 1993, featuring support with the Amiga 1200 for enhanced visuals, while almost all future conversions of the game were developed internally at Krisalis Software and each one features several changes and additions when compared with the original Amiga version. The first console port to be released was the Super Nintendo Entertainment System, which was first released in Japan by Yanoman Corporation before being published on western regions by Ocean Software in 1994. In North America, this port was renamed to The Adventures of Kid Kleets, and in Japan this port was renamed to Great Soccer Kid. In France, the Super Nintendo version was going to be released as a licensed game based on Moero! Top Striker, known in France as L'École des champions, however this version was never released. During the same year, it was ported to MS-DOS and only released in western regions. An Amiga CD32 was also developed and published by Krisalis in 1994 featuring mostly the same improved visuals when playing on Amiga 1200, a brand new level that was cut from the original version due to space constrains, as well as a new animated opening sequence created by a Japanese animation studio. A 3DO port developed by Team17 and published by Studio 3DO was first released in North America on 2 May 1994. In Japan, the 3DO port was released in Japan on 9 December 1994, using the name Great Soccer Kid like the Super Nintendo version. The port features redrawn artwork and sports a different GUI. A port of Soccer Kid for the Atari Jaguar was announced in early 1994 after Krisalis was signed to be a third-party developer by Atari Corporation for the system in September 1993, which was originally intended to be published by Ocean on a scheduled Q2 1995 release date, however it was never released during the official life span of the Jaguar on the market and went almost unreleased after Atari discontinued the system, until it was eventually picked up for release on 7 February 2000 by independent developer and publisher Songbird Productions. This version of the game features similar visuals to those of the original Amiga version, albeit with an improved color palette and a new UI. A Windows Mobile conversion of Soccer Kid was released worldwide on 4 April 2002. Several months later, the game was converted to the Game Boy Advance and published by Telegames after being showcased to the public at E3 2002, featuring enhanced visual effects but a lower resolution display. Krisalis later ported the GBA version to PlayStation and was only released in Europe by Telegames on 13 December 2003, becoming both the last version of the game to be officially released and the final version developed by the original team. A version of Soccer Kid for the Atari ST was being developed by Krisalis and was also previewed under its earlier title, Football Kid, by ST Review magazine on their September 1992 issue, however it never released on the platform for unknown reasons. A Sega Mega Drive version was reviewed by Spanish magazine Hobby Consolas, however the Mega Drive version was never released. Likewise, an Atari Lynx conversion of the game was also planned to be published by Telegames, but it was never released as well. Reception Amiga Power gave the Amiga version a score of 88% with a review by Stuart Campbell. He criticized the player character's moves as being too "fiddly": "you have to stop, line up your shot, get into trick shot mode and then actually do it. This all looks great and stylish and everything, but it can really slow the pace of the game down on occasion - sometimes when you're zipping along, you'd give anything just to be able to jump on a baddie's head and sort him out then-and-there without a lot of hanging around." However, he praised the graphics and particularly the "fearsome" difficulty, saying it gave the game good value for money. He compared the game to the acclaimed Arabian Nights, saying it was even better due to the superior control and greater focus on platform action. The 3DO version received a 6.2 out of 10 from Electronic Gaming Monthly. They praised the "technique" and "excellent graphics", but said the sound effects were lacking. GamePro gave it a wholly positive review, citing smooth controls, strong sound effects, good variety of musical tracks, and exceptional graphics. Reviewing the 3DO version, AllGame'''s Shawn Sackenheim wrote that Soccer Kid is "an impressive action/platformer that manages to be just different enough to warrant investigation."Next Generation'' reviewed the 3DO version of the game, rating it two stars out of five, and stated that "Hopefully, 3DO can evolve past this type of game faster than the systems of yesterday." Notes References External links Soccer Kid at AtariAge Soccer Kid at GameFAQs Soccer Kid at Giant Bomb Soccer Kid at MobyGames Soccer Kid at Soccer Kid 1993 video games 3DO Interactive Multiplayer games Amiga games Amiga 1200 games Atari Jaguar games Cancelled Atari Lynx games Cancelled Atari ST games Cancelled Sega Genesis games Amiga CD32 games DOS games Game Boy Advance games Krisalis Software games Ocean Software games Piko Interactive games Platform games PlayStation (console) games Songbird Productions games Sports fiction Super Nintendo Entertainment System games Telegames games Video games about children Video games about extraterrestrial life Video games developed in the United Kingdom Video games about pirates Video games set in England Video games set in Italy Video games set in Russia Video games set in Japan Video games set in the United States Yanoman games Single-player video games
45544472
https://en.wikipedia.org/wiki/DDR%20Holdings%20v.%20Hotels.com
DDR Holdings v. Hotels.com
DDR Holdings, LLC v. Hotels.com, L.P., , was the first United States Court of Appeals for the Federal Circuit decision to uphold the validity of computer-implemented patent claims after the Supreme Court's decision in Alice Corp. v. CLS Bank International. Both Alice and DDR Holdings are legal decisions relevant to the debate about whether software and business methods are patentable subject matter under Title 35 of the United States Code §101. The Federal Circuit applied the framework articulated in Alice to uphold the validity of the patents on webpage display technology at issue in DDR Holdings. In Alice, the Supreme Court held that a computer implementation of an abstract idea, which is not itself eligible for a patent, does not by itself transform that idea into something that is patent eligible. According to the Supreme Court, in order to be patent eligible, what is claimed must be more than the abstract idea. The implementation of the idea must be something beyond the "routine," "conventional" or "generic." In DDR Holdings, the Federal Circuit, applying the Alice analytical framework, upheld the validity of DDR's patent on its webpage display technology. Background DDR Holdings, LLC ("DDR") was formed by inventors Daniel D. Ross and D. Delano Ross, Jr. following the asset sale of their dot-com company, Nexchange (which was formed to utilize their invention). DDR filed a lawsuit against twelve entities including Hotels.com, National Leisure Group, World Travel Holdings, Digital River, Expedia, Travelocity.com, and Orbitz Worldwide for patent infringement. DDR settled with all but three of these defendants prior to an October 2012 jury trial in the United States District Court for the Eastern District of Texas. The jury found that neither of the patents at issue were invalid, that National Leisure Group, Inc. and World Travel Holdings, Inc. (collectively "NLG") directly infringed both these patents, that Digital River directly infringed one of the patents, and that DDR should be awarded $750,000 in damages. Following the verdict, the district court denied defendants’ motions for Judgment as a matter of law (JMOL) and entered final judgment in favor of DDR, consistent with the jury's findings. Defendants appealed, however, by the time of oral argument, DDR settled with Digital River, and Digital River's appeal was subsequently terminated. NLG continued its appeal. Patents-in-suit DDR is the assignee of U.S. Patent Nos. 7,818,399 ("the '399 patent") and 6,993,572 ("the '572 patent"), both of which are continuations of an earlier patent—U.S. Patent No. 6,629,135 ("the '135 patent"). The court's § 101 analysis focused on the '399 patent, entitled "Methods of expanding commercial opportunities for internet websites through coordinated offsite marketing." The Invention The '399 patent addresses a particular problem in the field of e-commerce when vendors advertise their products and services through a hosting page of an affiliate: [Vendors] are able to lure visitor traffic away from the affiliate. Once a visitor clicks on an affiliate ad and enters an online store, that visitor has left the affiliate's site and is gone. [This presents] a fundamental drawback of the affiliate programs--the loss of the visitor to the vendor. The '399 patent also identifies some other attempts at solving this problem: Affiliates are able to use "frames" to keep a shell of their own website around the vendor's site, but this is only a marginally effective solution. Some Internet affiliate sales vendors have begun placing "return to referring website" links on their order confirmation screens, an approach that is largely ineffective. This limitation of an affiliate program restricts participation to less trafficked websites that are unconcerned about losing visitors. Search engines and directories continue to increase in their usefulness and popularity, while banner ads and old-style links continue their rapid loss of effectiveness and popular usage. The '399 patent claims a process involving stored data concerning the visual elements responsible for the "look and feel" of the host website, where, upon clicking an ad for a third-party merchant's product, the customer is directed to a hybrid page generated by the host website that is a composite of the third-party merchant's product information and the look and feel elements of the host website. "For example, the generated composite web page may combine the logo, background color, and fonts of the host website with product information from the merchant." The Federal Circuit described the '399 patent as follows: The patents-in-suit disclose a system that provides a solution to this problem (for the host) by creating a new web page that permits a website visitor, in a sense, to be in two places at the same time… [T]he host website can display a third-party merchant’s products, but retain its visitor traffic by displaying this product information from within a generated web page that ‘gives the viewer of the page the impression that she is viewing pages served by the host’ website." [The system] instructs an Internet web server of an “outsource provider” to construct and serve to the visitor a new, hybrid web page that merges content associated with the products of the third-party merchant with the stored “visually perceptible elements” from the identified host website. A case note states that one way to accomplish the function would be with an <iframe> tag (or a <frameset> tag, which HTML5 allows but no longer supports). The court's opinions use the metaphor of the "store within a store" to describe what the invention does and how it works, stating that the inventor's idea was to have an Internet webpage that was like a warehouse store or department store (the dissenting opinion uses BJ's Wholesale Clubs as an illustrative example); other merchants' webpages would function like concessions or kiosks within the department store. A commentator asserted that the court's "store within a store" metaphor may not be the best way to look at this claimed invention, and that it may be more apt to characterize what the invention does as placing a frame around someone else's webpage and incorporating the frame and its content into the host's webpage. This is the effect that the invention accomplishes: Representative claim 19 of the '399 patent recites: "19. A system useful in an outsource provider serving web pages offering commercial opportunities, the system comprising: (a) a computer store containing data, for each of a plurality of first web pages, defining a plurality of visually perceptible elements, which visually perceptible elements correspond to the plurality of first web pages; (i) wherein each of the first web pages belongs to one of a plurality of web page owners; (ii) wherein each of the first web pages displays at least one active link associated with a commerce object associated with a buying opportunity of a selected one of a plurality of merchants; and (iii) wherein the selected merchant, the out-source provider, and the owner of the first web page displaying the associated link are each third parties with respect to one other; (b) a computer server at the outsource provider, which computer server is coupled to the computer store and programmed to: (i) receive from the web browser of a computer user a signal indicating activation of one of the links displayed by one of the first web pages; (ii) automatically identify as the source page the one of the first web pages on which the link has been activated; (iii) in response to identification of the source page, automatically retrieve the stored data corresponding to the source page; and (iv) using the data retrieved, automatically generate and transmit to the web browser a second web page that displays: (A) information associated with the commerce object associated with the link that has been activated, and (B) the plurality of visually perceptible elements visually corresponding to the source page." Legal landscape Under US law, all patentable inventions meet several general requirements: The claimed invention must be: statutory subject matter Judicial interpretation of this statute dictates that natural phenomena, laws of nature, and abstract ideas are not themselves patentable (although a particular application of a law of nature or an abstract idea might be patent-eligible). novel. nonobvious. useful. fully disclosed and enabled. As is entirely typical, the defendant argued that DDR's patents were invalid under all of these sections, but the primary litigation focus was on § 101 and whether DDR's patents were claiming an abstract idea which would not be patentable subject matter. In Alice, the Supreme Court clarified its two-prong framework, originally set forth in Mayo Collaborative Servs. v. Prometheus Labs., Inc., for evaluating the patent-eligibility of a claim under § 101. First one must determine whether the claim is directed to a patent-ineligible law of nature, natural phenomenon, or abstract idea. If so, then one determines whether any additional claim elements transform the claim into a patent-eligible application that amounts to significantly more than the ineligible concept itself. Under Alice, Mayo, and Ultramercial, claims are patent-ineligible under § 101 if they are directed to patent-ineligible subject matter (i.e. abstract ideas, laws of nature, and natural phenomena) and do not contain an inventive concept that sufficiently transforms the claim into an application of the underlying idea that restricts the claim to something significantly different from the ineligible subject matter it is directed to. This embodiment must be something more than typical operations performed on a generic computer. Following the Alice decision, several cases invalidated patents covering computer-implemented inventions as ineligible abstract ideas, including Ultramercial. Because the analyses in these decisions are somewhat ambiguous (on, e.g., defining the scope and standard of the term "abstract idea"), many inventors, bloggers, scholars, and patent lawyers have struggled with determining their full implication, especially as they relate to software claims, and some have even questioned the patentability of computer-implemented inventions in general. Decision Judge Chen authored the opinion of the Federal Circuit, joined by Judge Wallach, which invalidated DDR's ‘572 patent as anticipated (overruling the District Court) and affirmed the District Court’s denial of NLG’s renewed motions for JMOL on invalidity and noninfringement of the ‘399 patent. The Federal Circuit Court held that the relevant claims of the ‘399 patent were directed to patent eligible subject matter and that the jury was presented with substantial evidence on which to base its finding that NLG infringed the ‘399 patent. Judge Mayer authored a dissenting opinion, arguing that the ‘399 patent was "long on obfuscation but short on substance[,]" and criticized the invention as "so rudimentary that it borders on the comical." He interpreted Alice to create a "technological arts" test which DDR's claims failed because they were directed toward an entrepreneurial objective (i.e. "retaining control over the attention of the customer") rather than a technological goal. § 101 Analysis The case is most significant for its discussion of 35 U.S.C. § 101 and the concept of an unpatentable abstract idea as it applies to software and business methods. In this discussion, the Federal Circuit applied the two-step test for patentability set forth in Alice to determine that DDR's '399 patent claims are directed to patent-eligible subject matter. First, it considered whether the claims were directed to a patent-ineligible abstract idea. Judge Chen does not arrive at a clear answer to this inquiry. Instead, he opts to ground his opinion in the more perceptible nature of eligibility should the analysis proceed to step two, without deciding whether that step is actually necessary. Step 1: Abstract idea? As an initial matter, the court must determine whether the claims at issue are directed to a patent ineligible concept (e.g. an abstract idea). At this step, the court observed that distinguishing between a patentable invention and an abstract idea "can be difficult, as the line separating the two is not always clear." Judge Chen acknowledged that the invention could be characterized as an abstract idea, such as "making two e-commerce web pages look alike," but also noted that the asserted claims of the ‘399 patent "do not recite a mathematical algorithm . . . [n]or do they recite a fundamental economic or longstanding commercial practice." It reviewed several Supreme Court cases useful in identifying claims directed to abstract ideas. However, the Federal Circuit Court never offered a precise definition of an unpatentable "abstract idea" nor did it explicitly decide whether the '399 claims are directed to such ineligible subject matter. Instead, the court concludes that, even stipulating any of the characterizations of the alleged abstract idea put forth by defense counsel and the dissent, the '399 claims still contain an inventive concept sufficient to render them patent-eligible under step two of the Alice analysis. Step 2: Inventive concept In step two, the court must "consider the elements of each claim — both individually and as an ordered combination — to determine whether the additional elements transform the nature of the claim into a patent-eligible application of that abstract idea. This second step is the search for an ‘inventive concept,’ or some element or combination of elements sufficient to ensure that the claim in practice amounts to ‘significantly more’ than a patent on an ineligible concept." In spite of the business-related nature of the claims (retaining or increasing website traffic) and the fact that they could be implemented on a generic computer, the court highlighted that the claims did not simply take an abstract business method from the pre-internet world and implement it on a computer. Instead, the claims addressed a technological problem "particular to the internet" by implementing a solution specific to that technological environment and different from the manner suggested by routine or conventional use within the field. The majority opinion characterized the problem as "the ephemeral nature of an Internet 'location' [and] the near-instantaneous transport between these locations made possible by standard communication protocols. The majority distinguished this problem, which they found was "particular to the Internet," from the circumstances inherent in the "store within a store" schemes—in traditional "brick and mortar" warehouse stores with cruise vacation package kiosks, visitors to the kiosk are still inside the warehouse store when making their kiosk purchases. Judge Chen thus found that "the claimed solution is necessarily rooted in computer technology in order to overcome a problem specifically arising in the realm of computer networks." The DDR court differentiated the claims of the ‘399 patent from those that "merely recite the performance of some business practice known from the pre-Internet world along with the requirement to perform it on the Internet." Instead, the court explained, the claims of patent ‘399 "address the problem of retaining website visitors that, if adhering to the routine, conventional functioning of Internet hyperlink protocol, would be instantly transported away from a host's website after clicking on an advertisement and activating a hyperlink." Because the invention "overrides the routine and conventional sequence of events ordinarily triggered by the click of a hyperlink," it did not employ mere ordinary use of a computer or the Internet. Further, the court held, the claims included additional features that limit their scope to not preempt every application of any of the abstract ideas suggested by NLG. Viewed individually and as an ordered combination, the DDR court concluded that the claims these aspects of the invention established an "inventive concept" for resolving an Internet-centric problem and were therefore directed to patent-eligible subject matter. Distinctions from patent-ineligible claims of past cases The court found that the ‘399 patent claims were significantly different from the patent-ineligible claims in Alice, Ultramercial, buySAFE, Accenture and Bancorp, in that the ‘399 claims did not "(1) recite a commonplace business method aimed at processing business information, (2) apply a known business process to the particular technological environment of the Internet, or (3) create or alter contractual relations using generic computer functions and conventional network operations." Unlike other cases recently decided under the Alice framework, the DDR court stated that the ‘399 patent does not "broadly and generically claim use of the Internet to perform an abstract business practice (with insignificant added activity)." Instead, the claims "specify how interactions with the Internet are manipulated to yield a desired result -- a result that overrides the routine and conventional sequence of events ordinarily triggered by the click of a hyperlink." The claimed system changes the normal operation of the Internet so that the visitor is directed to a "hybrid web page that presents product information from the third-party and visual 'look and feel' elements from the host website." Thus, Judge Chen concluded that the claimed invention "is not merely the routine or conventional use of the Internet." In Ultramercial, Inc. v. Hulu, the court analyzed the eligibility of patent claims directed to a method for distributing copyrighted media products over the Internet where the consumer receives the content, paid for by an advertiser, in exchange for viewing an advertisement. Although the problem solved by the invention arguably was particular to the Internet, the court concluded that the steps of the claims are "an abstraction – an idea, having no particular concrete or tangible form." The court went on to hold that the limitations of the claims "do not transform the abstract idea ... into patent-eligible subject matter because the claims simply instruct the practitioner to implement the abstract idea with routine, conventional activity." The Federal Circuit's DDR decision distinguished itself from Ultramercial by noting: "The ’399 patent’s claims are different enough in substance from those in Ultramercial because they do not broadly and generically claim "use of the Internet" to perform an abstract business practice (with insignificant added activity). Unlike the claims in Ultramercial, the claims at issue here specify how interactions with the Internet are manipulated to yield a desired result—a result that overrides the routine and conventional sequence of events ordinarily triggered by the click of a hyperlink… When the limitations of the ’399 patent’s asserted claims are taken together as an ordered combination, the claims recite an invention that is not merely the routine or conventional use of the Internet. It is also clear that the claims at issue do not attempt to preempt every application of the idea of increasing sales by making two web pages look the same, or of any other variant suggested by NLG. Rather, they recite a specific way to automate the creation of a composite web page by an "outsource provider" that incorporates elements from multiple sources in order to solve a problem faced by websites on the Internet." The majority also distinguished the brick-and-mortar analog to the patented invention—in-store kiosks—as "not hav[ing] to account for the ephemeral nature of an Internet ‘location’ or the near-instantaneous transport between these locations made possible by standard Internet communication protocols, which introduces a problem that does not arise in the ‘brick and mortar’ context." Dissent Judge Mayer in his dissent proposed a brick and mortar analog to the claimed invention where an individual shop within a larger store had the same décor as the larger shop to "dupe" shopper to believing he/she was in the larger store. Hence, the claimed invention did not address a unique problem to the internet. Significance and reception Commentary Professor Crouch, in the Patently-O blog, commented: "The case is close enough to the line that I expect a strong push for en banc review and certiorari. Although Judge Chen’s analysis is admirable, I cannot see it standing up to Supreme Court review and, the holding here is in dreadful tension with the Federal Circuit’s recent Ultramercial decision." Gene Quinn, a patent lawyer and blogger at , doubts that this case can be reconciled with Ultramercial, despite the Federal Court's attempt to distinguish the two. Quinn found the difference between DDR and Ultramercial "thin" and one "that is not at all likely to lead to a repeatable and consistent test that can be applied in a predictable way." Michael Borella, a patent attorney, said in the blog: "Not only does this case give us another data point of how a computer-implemented invention that incorporates an abstract idea can be patent-eligible (Diamond v. Diehr is the other notable example), but it also provides the first appellate use of the second prong of the Alice test to do so." District courts As one professorial commentator noted, "Because DDR Holdings is the only post-Alice Federal Circuit decision so far to uphold a patent against a § 101 challenge, patentees have been quick to cite it and accused infringers have found ways to distinguish it." Among the district court cases interpreting and applying DDR Holdings are: KomBea Corp. v. Noguar L.C., 2014 U.S. Dist. LEXIS 177186 (D. Utah Dec. 23, 2014) – "The patents-in-suit are distinguishable from the patents in DDR Holdings. First, the patents-in-suit are not directed toward solving a new problem, unique to a technological field. Rather, the patents-in-suit are directed toward performing fundamental commercial practices more efficiently. Second, the patents-in-suit are not a new solution to a unique problem; they only employ a combination of sales techniques and basic telemarketing technology to create an efficient system. . . . In this case, the fact that Defendant’s patents-in-suit are directed toward abstract ideas that are more efficiently executed with the use of a generic computer does not make the patents eligible for protection. Therefore, the Court finds that the claims, individually and collectively, do not transform the abstract ideas within the claims into an inventive concept. As such, the patents-in-suit fail to transform the abstract ideas they claim into patent-eligible subject matter." MyMedicalRecords, Inc. v. Walgreen Co., 2014 U.S. Dist. LEXIS 176891 (C.D. Cal. Dec. 23, 2014) – "Unlike the claims in DDR these claims are directed to nothing more than the performance of a long-known abstract idea 'from the pre-Internet world' — collecting, accessing, and managing health records in a secure and private manner — on the Internet or using a conventional computer. The patent claims are not 'rooted in computer technology in order to overcome a problem specifically arising in the realm of computer networks' [as in] DDR. Rather, the patent recites an invention that is merely the routine and conventional use of the Internet and computer with no additional specific features." Open TV, Inc. v. Apple, Inc., – F. Supp. 2d – (N.D. Cal. Apr. 6, 2015) – "Even construing the claims as Open TV does, the invention claimed by the `799 Patent is not 'necessarily rooted in computer technology in order to overcome a problem specifically arising in the realm of [television] networks.' DDR Holdings, 773 F.3d at 1257. Rather, as noted above, the problem of transmitting confidential information using unsecure communication methods has existed for centuries, long before the advent of interactive television networks. . . . Here, as described more fully above in the Court's analysis of the first prong of the Alice test, the `799 Patent does not claim a solution to a problem that arose uniquely in the context of interactive television networks. Furthermore, the `799 Patent claims recite a method that does not go beyond the 'routine or conventional use' of existing electronic components." Messaging Gateway Solns. LLC v. Amdocs, Inc., – F. Supp. 2d – (D. Del. Apr. 15, 2015) – "The Court finds that Claim 20 contains an inventive concept sufficient to render it patent-eligible. Like the claims in DDR Holdings, Claim 20 'is necessarily rooted in computer technology in order to overcome a problem specifically arising in the realm of computer networks.' See id. at 1257. Claim 20 is directed to a problem unique to text-message telecommunication between a mobile device and a computer. The solution it provides is tethered to the technology that created the problem." Intellectual Ventures I LLC v. Symantec Corp., 100 F. Supp. 3d 371 (D. Del. 2015), reversed in pertinent part, 838 F.3d 1307 (Fed. Cir. 2016) – "The Federal Circuit in DDR Holdings, 773 F.3d at 1259, held that the 'Internet-centric' claims at issue there were patent eligible. Claim 7 of the '610 patent is 'Internet-centric.' In fact, the key idea of the patent is that virus detection can take place remotely between two entities in a telephone network. This is advantageous because it saves resources on the local caller and calling machines and more efficiently executes virus detection at a centralized location in the telephone network. Claims that 'purport to improve the functioning of the computer itself' or 'effect an improvement in any other technology or technical field' may be patentable under § 101. Alice." See also Bilski v. Kappos References External links United States patent case law Software patent case law United States Court of Appeals for the Federal Circuit cases 2015 in United States case law Expedia Group
5984746
https://en.wikipedia.org/wiki/Card%20sharing
Card sharing
Card sharing, also known as control word sharing, is a method of allowing multiple clients or digital television receivers to access a subscription television network with only one valid subscription card. This is achieved by electronically sharing a part of the legitimate conditional access smart card's output data, enabling all recipients to gain simultaneous access to scrambled DVB streams, held on the encrypted television network. Typically, a legitimate smart card is attached to a host digital television receiver, which is equipped with software to share the decrypted 64-bit "control word" key over a computer network, such as the Internet. Once a client receives this key, they can decrypt the encrypted content as though they were using their own subscription card. The theory of card sharing as a potential attack vector on pay TV encryption was provided in the book European Scrambling Systems by John McCormac in 1996; leading to the term "McCormac Hack" to be used to describe early card-sharing systems. Basic operation The security of conditional access smart card technology is limited by the security of the DVB standard in which it operates. The standard practice of a legitimate smart card is to decrypt an ECM (Entitlement Control Message), which then provides the control word, which allows the viewing of scrambled material. With card sharing, however, the smart card and its security features are bypassed; software intercepts the decrypted control word and allows the user to share it across a computer network. Pirate decryption Card sharing has established itself as popular method of pirate decryption. Much of the development of card sharing hardware and software has taken place in Europe, where national boundaries mean that home users are able to receive satellite television signals from many countries but are unable to legally subscribe to them due to licensing restrictions on broadcasters. Because the length of the complete control word is so small (64 bits), delivery of the control words to many different clients is easily possible on a home internet connection. This has sparked the creation of sharing network groups, in which users can access the group by sharing their subscription cards with the group, and in turn, being capable of receiving the channels which all users' cards can decrypt, as though the user owned every single subscription card connected to the network. Other networks have also been created, whereby one server has multiple legitimate subscription cards connected to it. Access to this server is then restricted to those who pay the server's owner their own subscription fee. Multiple receiver use An arguably legitimate use for card sharing is the sharing of the control word within a home network, where the subscriber is authorised by the subscription television network to decrypt their signal, using one smart card. Content providers usually provide means for viewing channels on a second smart card, provided at extra cost. An example of this is Sky Multiroom, used in the United Kingdom. However, in some cases the contract between the subscriber and the content provider implicitly or explicitly prohibits this kind of card sharing. Countermeasures Card sharing is a particular concern to conditional access providers, and their respective pay-TV companies, as well as the DVB consortium. Card sharing utilises the integral scrambling system of the DVB standard, DVB-CSA, meaning that every provider of scrambled DVB content has potential to be affected by it. In response, several counter measures have been implemented by various parties, with the aim of permanently preventing it. Technical One technical method, implemented by providers such as Irdeto and NDS, is to update the software of digital receivers provided by the subscription television service. This software implements a further decryption layer, held within the receiver. Rather than sending a plain text control word from the smart card to the receiver's microprocessor, which can be intercepted, the decrypted ECM will in fact be an encrypted control word, which can only be decrypted by a legitimate, non card sharing capable, receiver. A simpler method, used by several providers, is to simply increase the frequency of control word changes. With changes occurring as frequently as once every few seconds, extra stress is put onto the smart card sharing system, meaning that clients may be frustrated by short, frequent, missed viewing periods. Another method gaining traction is "Card pairing". Effectively linking the card with the connected device (e.g the set-top box, CAM module or TV's decoding chipset) and locking the card to be used only on that device. Legal One significant problem for internet card sharing is that the activity can be traced, through tracing of the originating IP address - though this does not identify an individual it can may to a specific ISP and/or TV service subscription. Cases of criminal charges have been brought against card sharing hosts. In August 2009, a Liverpool man was prosecuted for re-selling the services of a card sharing network to paying customers. It was the first such case of its type in the UK. Two further raids occurred in the UK during early 2011, instigated by UK Cable Operator Virgin Media. DVB-CSA3 In 2007, the DVB Project approved and began licensing a new scrambling system, CSA3, for protecting DVB content. This new system, upon implementation, will attempt to eradicate many of the flaws with the original DVB-CSA system, including introducing an AES based 128-bit key system. The system will reportedly be "hardware friendly and software unfriendly", indicating that reverse engineering of the system, required for the creation of card sharing applications, will be very difficult. See also Smart card Pirate decryption References Digital television Cable television technology
1022571
https://en.wikipedia.org/wiki/Digital%20Negative
Digital Negative
Digital Negative (DNG) is a patented, open, lossless raw image format developed by Adobe and used for digital photography. Adobe's license allows use without cost on the condition that the licensee prominently displays text saying it is licensed from Adobe in source and documentation, and that the license may be revoked if the licensee brings any patent action against Adobe or its affiliates related to the reading or writing of files that comply with the DNG Specification. It was launched on September 27, 2004. The launch was accompanied by the first version of the DNG specification, plus various products, including a free-of-charge DNG converter utility. All Adobe photo manipulation software (such as Adobe Photoshop and Adobe Lightroom) released since the launch supports DNG. DNG is based on the TIFF/EP standard format, and mandates significant use of metadata. Use of the file format is royalty-free; Adobe has published a license allowing anyone to exploit DNG, and has also stated that there are no known intellectual property encumbrances or license requirements for DNG. Adobe stated that if there were a consensus that DNG should be controlled by a standards body, they were open to the idea. Adobe has submitted DNG to ISO for incorporation into their revision of TIFF/EP. Rationale for DNG Adobe states that, given the existence of a wide variety of camera-brand-specific raw image formats, it introduced DNG as a standardized and backward-compatible universal file format. It is based on the TIFF 6.0 standard. Various professional archivists and conservationists, working in institutional settings have adopted DNG for archival purposes. Objectives These objectives are repeatedly emphasized in Adobe documents: Digital image preservation (sometimes known as "archiving"): to be suitable for the purpose of preserving digital images as an authentic resource for future generations. Assessment: The US Library of Congress states that DNG is a recommended alternative to other raw image formats: "Less desirable file formats: RAW; Suggested alternatives: TIFF_UNC, JP2_J2K_C_LL, DNG". The Digital Photography Best Practices and Workflow (dpBestflow) project, funded by the United States Library of Congress and run by the American Society of Media Photographers (ASMP), singles out DNG, and states "DNG files have proven to be significantly more useful than the proprietary raw files in our workflow". Easy and/or comprehensive exploitation by software developers: to enable software to be developed without the need for reverse engineering; and to avoid the need for frequent software upgrades and re-releases to cater for new cameras. Assessment: Software could support raw formats without DNG, by using reverse engineering and/or dcraw; DNG makes it easier, and many software products can handle, via DNG, images from cameras that they have no specific knowledge of. An unresolved restriction is that any edit/development settings stored in the DNG file by a software product are unlikely to be recognized by a product from a different company. (This problem is not specific to DNG). In-camera use by camera manufacturers: to be suitable for many camera manufacturers to use as a native or optional raw image format in many cameras. Assessment: About 12 camera manufacturers have used DNG in-camera. About 38 camera models have used DNG. Raw image formats for more than 230 camera models can be converted to DNG. Multi-vendor interoperability: to be suitable for workflows where different hardware and software components share raw image files and/or transmit and receive them. Characteristics All of the above objectives are facilitated or enabled by most of these characteristics: Freely-available specification: this can be downloaded from the Adobe website without negotiation or needing justification. Format based on open specifications and/or standards: DNG is compatible with TIFF/EP, and various open formats and/or standards are used, including Exif metadata, XMP metadata, IPTC metadata, CIE XYZ coordinates and JPEG. Self-contained file format: a DNG file contains the data (raw image data and metadata) needed to render an image without needing additional knowledge of the characteristics of the camera. Version control scheme: it has a version scheme built into it that allows the DNG specification, DNG writers, and DNG readers to evolve at their own paces. Freely-available source-code-based software development kit (SDK): there are three aspects - there is an SDK; it is source-code-based (as can be verified by examination); and it can be downloaded from the Adobe website without needing justification. Documented to have no known intellectual property encumbrances or license requirements: there is both a "Digital Negative (DNG) Specification Patent License" which says that anyone can exploit DNG, and a statement that there are no known intellectual property encumbrances or license requirements for DNG. Lossless and lossy compression (optional): DNG support optional lossless and (since version 1.4) also lossy compression. The lossy compression losses are practically imperceptible in real world images. Technical summary A DNG file always contains data for one main image, plus metadata, and optionally contains at least one JPEG preview. It normally has the extension "dng" or "DNG". DNG conforms to TIFF/EP and is structured according to TIFF. DNG supports various formats of metadata (including Exif metadata, XMP metadata, IPTC metadata) and specifies a set of mandated metadata. DNG is both a raw image format and a format that supports "non-raw", or partly processed, images. The latter (non-raw) format is known as "Linear DNG". Linear DNG is still scene-referred and can still benefit from many of the operations typically performed by a raw converter, such as white balance, the application of a camera color profile, HDR compositing, etc. All images that can be supported as raw images can also be supported as Linear DNG. Images from the Foveon X3 sensor or similar, hence especially Sigma cameras, can only be supported as Linear DNG. DNG can contain raw image data from sensors with various configurations of color filter array (CFA). These include: conventional Bayer filters, using three colors and rectangular pixels; four-color CFAs, for example the RGBE filter used in the Sony Cyber-shot DSC-F828; rectangular (non-square) pixels, for example as used in the Nikon D1X; and offset sensors (for example with octagonal pixels) such as Super CCD sensors of various types, as used in various Fujifilm cameras. (Or combinations of these if necessary). DNG specifies metadata describing these individual parameters; this is one significant extension to TIFF/EP. When used in a CinemaDNG movie clip, each frame is encoded using the above DNG image format. The clip's image stream can then be stored in one of two formats: either as video essence using frame-based wrapping in an MXF file, or as a sequence of DNG files in a specified file directory. Contrary to its name (Digital Negative) the DNG format doesn't distinguish negative and positive data - all data is considered to be describing a positive image. While this is not an issue when working with images from digital cameras (which are always positive), working with scanned (by a film scanner or DSLR copy stand) film negatives saved as raw DNG files is complicated, because the resultant image is not automatically inverted and thus impossible to be used directly. A way to get around this is using an inverted curve in the photo editing application, however this reverses the effect of the image controls (Exposure, Shadow and Highlight details, etc.) which complicates the photo editing. Timeline This provides a mixture of the dates of significant events (such as "the first X") and various counts of usage at the anniversaries of the launch (each 27 September). Counts of products and companies that use DNG in some way are provided primarily for illustration. They are approximate, and include products that are no longer sold. The purpose is mainly to demonstrate that such products and companies exist, and to show trends. Convertible raw image formats (camera models whose raw images can be converted to DNG) only include official support by Adobe DNG converters; not unofficial support by Adobe products (sometimes reaching about 30), nor support by other DNG converters. 2003, late: Adobe started work on the DNG specification. 2004, early: Adobe started talking to other companies about DNG. Launch, 2004, 27 September: Adobe launched DNG. Specification version 1.0.0.0 published. Convertible raw image formats: 60+ 2005, February: Specification version 1.1.0.0 published. 2005, June: First digital back to write DNG, the Leica DMR (Digital Modul R) back for the R8 and R9. 2005, July: First camera manufacturer to provide a DNG converter for its own raw file formats - Hasselblad's FlexColor. 2005, July: First genuine digital SLR camera to write DNG, Hasselblad H2D. 1st anniversary, 2005, 27 September: Camera manufacturers: 4. Camera models: 7. Software products: 70+. Convertible raw image formats: 70+. 2005, October: First compact camera to write DNG, Ricoh GR Digital. 2006, July: First monochrome digital back to write DNG, MegaVision E Series MonoChrome. 2006, September: First rangefinder camera to write DNG, Leica M8. 2006, September: First camera to offer the user a choice of proprietary raw or DNG, Pentax K10D. 2nd anniversary, 2006, 27 September: Camera manufacturers: 8. Camera models: 9. Software products: 120+. Convertible raw image formats: 110+. 2007, May: First raw converter & photo-editor whose first raw-handling release only supported DNG, Pixelmator. 2007, July: First underwater camera to write DNG, Sea&Sea DX-1G. (Based on Ricoh Caplio GX100). 3rd anniversary, 2007, 27 September: Camera manufacturers: 10. Camera models: 13. Software products: 170+. Convertible raw image formats: 160+. 2007, October: First digital scan back system and first 360-degree panorama system to write DNG, Seitz 6x17 Digital and Seitz Roundshot D3 with D3 digital scan back. 2008, February: First software on a mobile phone to write DNG, Tea Vui Huang's "DNG Phone Camera" for Nokia. 2008, April: Adobe announces CinemaDNG initiative, using DNG as the basis for the individual raw images of a movie. 2008, May: Specification version 1.2.0.0 published. 2008, September: First movie camera to use DNG as a raw image format, Ikonoskop A-cam dII. 4th anniversary, 2008, 27 September: Camera manufacturers: 13. Camera models: 29. Software products: 200+. Convertible raw image formats: 180+. 2008, September: First DNG converter running on Linux (among several other things), digiKam. 2009, spring/summer: First digiscope with built-in camera to write DNG, Zeiss Photoscope 85 T* FL. 2009, June: Specification version 1.3.0.0 published. 5th anniversary, 2009, 27 September: Camera manufacturers: 14. Camera models: 38. Software products: 220+. Convertible raw image formats: 230+. 2009, November: First "interchangeable unit" camera to write DNG, Ricoh GXR. 2010, February: First 3D movie camera to write DNG, Ikonoskop A-cam3D. 2010, March: First medium format camera to offer the user a choice of proprietary raw or DNG, Pentax 645D. 6th anniversary, 2010, 27 September: Camera manufacturers: 14. Camera models: 47. Software products: 240+. Convertible raw image formats: 290+. 2012, September: Specification version 1.4.0.0 published. 2019, May: Specification version 1.5.0.0 published. During the first 5 years when about 38 camera models were launched that wrote DNG, Adobe software added support for about 21 Canon models, about 20 Nikon models, and about 22 Olympus models. Reception The reaction to DNG has been mixed. A few camera manufacturers stated their intention to use DNG at launch. They first supported DNG about 9 months after launch. Several more niche and minority camera manufacturers added support after this (e.g. Leica). The largest camera manufacturers have apparently never indicated an intention to use DNG (e.g. Nikon and Canon). Some software products supported DNG within 5 months of launch, with many more following. Some only support DNG from cameras writing DNG, or from cameras supported via native raw image formats. OpenRAW was an advocacy and lobby group with the motto "Digital Image Preservation Through Open Documentation". They became opposed to DNG. Some photographic competitions do not accept converted files, and some do not accept DNG. DNG conversion "DNG conversion" refers to the process of generating a DNG file from a non-DNG image. (This is in contrast to "raw conversion", which typically refers to reading and processing a raw file, which might be a DNG file, and generating some other type of output from it). DNG conversion is one of the sources of DNG files, the other being direct output from cameras and digital backs. Several software products are able to do DNG conversion. The original such product is Adobe DNG Converter or DNG Converter, a freely-available stand-alone utility from Adobe. Other Adobe products such as the ACR plugin to Photoshop or Lightroom can also generate DNG files from other image files. Most DNG converters are supplied by companies other than Adobe. For example: The software that Pentax supplies with all their dSLR cameras can convert PEF raw image files from Pentax cameras into DNG files. Flexcolor and Phocus from Hasselblad can convert 3FR raw image files from Hasselblad cameras and digital backs into DNG files. Capture One from Phase One is a raw converter that can process not only raw image files from Phase One digital backs, but also raw image files from many other cameras too. Capture One can save images from many of those cameras to DNG. KDE Image Plugin Interface is an API that can save the images it is processing to DNG. It can be used standalone or with any image processing applications of the KDE desktop environment under Linux and Windows. A number of DNG converters have been developed by "amateurs" to enable raw images from their favored camera or digital back to be processed in a large range of raw converters. These include cases where cameras have been hacked to output raw images that have then been converted to DNG. The process of DNG conversion involves extracting raw image data from the source file and assembling it according to the DNG specification into the required TIFF format. This optionally involves compressing it. Metadata as defined in the DNG specification is also put into that TIFF assembly. Some of this metadata is based on the characteristics of the camera, and especially of its sensor. Other metadata may be image-dependent or camera-setting dependent. A DNG converter must therefore have knowledge of the camera model concerned, and be able to process the source raw image file including key metadata. Optionally a JPEG preview is obtained and added. Finally, all of this is written as a DNG file. DNG conversion typically leaves the original raw image file intact. For safety, many photographers retain the original raw image file on one medium while using the DNG file on another, enabling them to recover from a range of hardware, software and human failures and errors. For example, it has been reported in user forums that some versions of the Adobe DNG Converter don't preserve all the raw data from raw images from some camera models. Summary of products that support DNG in some way This section summarizes other more comprehensive lists. Adobe products All raw image file handling products from Adobe now support DNG. Adobe DNG Converter was utility software published by Adobe Systems on September 27, 2004. It converts different camera raw format files into the Digital Negative (DNG) standard. It also supports lossless data compression when converting. The program is free of charge. It can be downloaded at Adobe's site (for Microsoft Windows and the Apple Macintosh). Digital cameras and related software Use by camera manufacturers varies; there are about 15 camera manufacturers that use DNG, including a few that specialize in movie cameras: Modern smartphone cameras using iOS 10, Android 5 or Windows 10 Mobile that are capable of shooting RAW images typically use DNG (e.g. Samsung and Nokia). High-end Nokia (now Microsoft) Lumia smartphones like Nokia Lumia Icon, 930, 950, 1020 and 1520, were the first smartphone cameras to support DNG files. Samsung's high-end smartphones (Galaxy S6 or later) also uses DNG. Huawei Android phones using Leica-designed cameras (such as the Mate 9 and Mate 10 Pro) can save raw files to DNG format when shooting in "Pro" mode. Niche camera manufacturers typically use DNG in new cameras (including a digiscope, panorama cameras, and at least one movie camera). The article on raw image formats illustrates the complicated relationship between new raw image formats and third-party software developers. Using DNG provides immediate support for these cameras by a large range of software products. Some low market share but conventional camera manufacturers use DNG in new cameras. Camera manufacturers that do not supply their own software for processing raw images typically, but not always, use DNG. Pentax typically offers users the option of whether to use Pentax's own raw image format (PEF) or DNG, but some, for example Pentax Q, Q10 and Q7, do not support PEF. For example, the digital SLR Camera Pentax K-x does offer the ability to save PEF or DNG or even DNG+ which saves two files, a DNG and a separate JPEG file at the same time. If a camera uses DNG, and that camera manufacturer supplies software, it will support DNG. It may support DNG only from their own cameras, or support it more generically. Canon, Nikon, Sony, Panasonic, Olympus and Fuji do not use DNG in their cameras. If a camera manufacturer's cameras do not use DNG, its software is unlikely to support DNG unless that software is also sold independently of the cameras. Some digital cameras that support DNG: Casio supports DNG in their Exilim PRO EX-F1 and Exilim EX-FH25. DxO supports DNG in their DxO ONE camera (introduced 2015). Leica's Digital Modul R for the Leica R8 or Leica R9 and the Leica M8 or Leica M9 natively support the DNG format. MegaVision E Series Monochrome back. Panoscan MK-3 digital panoramic camera. Pentax supports DNG in their 645D, 645Z, K10D, K20D, K200D, K2000, K-7, K-x, K-r, K-5, K-30, K-5II(s), K-50, K-500, K-3, K-3II, K-S1, K-S2, K-70 and K-1 DSLR cameras; alongside the K-01, Q, Q10, Q7 and Q-S1 mirrorless cameras. Ricoh supports DNG in the Ricoh Digital GR, considered a professional compact, and the Ricoh Caplio GX. Ricoh GXR mirrorless interchangeable lens camera unit use also DNG. Samsung supports DNG in their Pro815 "prosumer" camera and GX-10 and GX-20 DSLR cameras. Sea&Sea DX‐1G underwater camera. Seitz Roundshot D3 digital back, used in cameras such as the 6×17. Sigma FP Native support, SD Quattro cameras support DNG but may require a firmware update. Silicon Imaging Silicon Imaging Digital Cinema SI-1920HDVR. Sinar now uses DNG as the raw file standard for their eMotion series of digital backs. Kron Technologies supports DNG in their Chronos 1.4 and Chronos 2.1-HD high speed cameras. Apple's iPhones and other iOS devices expose an API for third-party apps such as Halide or Lightroom CC to capture DNG images. The native Camera app processes to JPEG or HEIF by default. Starting with iOS 14.3 on the iPhone 12 Pro/Pro Max or later, Apple's ProRAW feature can be enabled which allows the Camera app to capture and save as DNG images. ProRAW combines the RAW image with the iPhone's image processing information into a 12-bit DNG file. Some of the Canon cameras can shoot as DNG using additional free software CHDK. The built-in camera function in the Adobe Lightroom Mobile app saves to DNG format. DJI supports DNG in its middle-end and high-end drones. Third-party software Support by software suppliers varies; there are of the order of 200 software products that use DNG. The majority of raw handling software products support DNG. Most provide generic support, while a few support it only if it is output directly from a camera. The type of support varies considerably. There appear to be very few third-party software products that process raw images but don't support DNG. This may reflect the difficulty of discovering all of those that do not. Versions of the specification All versions of the specification remain valid, in the sense that DNG files conforming to old versions should still be read and processed by DNG readers capable of processing later versions. DNG has a version scheme built into it that allows the DNG specification, DNG writers, and DNG readers, to evolve at their own paces. Each version of the specification describes its compatibility with previous versions. 1.0.0.0, published September 2004 This version accompanied the launch of DNG and related products. It was a rare, possibly unique, example of a raw image format specification published by its owner. It was adequate for representing typical images, but it had a few errors and deficiencies (specifically the lack of support of "masked pixels" and an inadvertent deviation from the JPEG specification) that required it soon to be replaced by the next version. 1.1.0.0, published February 2005 This version corrected the flaws in the first version. It has proved capable of representing raw images for a large variety of cameras (both when written in-camera or via conversion from other raw image formats) for a few years, and it is the version still typically written in-camera. 1.2.0.0, published May 2008 This version was based on experience and feedback from other companies about DNG since its launch. It introduced many new features, especially several new options for color specification under the general heading of "Camera Profiles". These are mainly of value to software products wanting their own flavor of color handling. This version permits administrative control of Camera Profiles, including calibration signatures and copyright information. 1.3.0.0, published June 2009 This version added various improvements, but the major change was to introduce "Opcodes". An Opcode is an algorithm to be applied to some or all of the image data, described in the specification, and (optionally) implemented in the product that reads and processes the DNG file. The DNG file itself holds lists of Opcodes to be executed, together with the parameters to be applied on execution. In effect, the DNG file can contain lists of "function calls" to be executed at various stages in the raw conversion process. For example, the WarpRectilinear Opcode "applies a warp to an image and can be used to correct geometric distortion and lateral (transverse) chromatic aberration for rectilinear lenses". This is an example of an algorithm that cannot be applied to the raw image data itself before it is placed into the DNG file, because it should be executed after demosaicing. There are 13 Opcodes described in this version, and each Opcode is accompanied by a specification version so that more can be added in future. 1.4.0.0, published September 2012 This version added Floating Point Image Data, Transparent Pixels, Proxy DNG Files, and additional tags. It also added SampleFormat and Predictor. 1.5.0.0, published May 2019 This version added Depth Maps and processed Enhanced Image Data that can be stored alongside the raw image, with some details of the processing performed. 1.6.0.0, published December 2020 This version added demosasic information, Semantic Maps and more. CinemaDNG, published September 2009 CinemaDNG uses DNG for each frame of a movie clip. There are additional tags specifically for movies: TimeCodes and FrameRate. It is not clear whether these tags will be added to a later version of the DNG specification, or will remain separately described only in the CinemaDNG specification. Standardization DNG is not (yet) a standard format, but is based on several open formats or standards and is being used by ISO in its revision of TIFF/EP. A timeline: 2001: The ISO standard raw image format, ISO 12234-2, better known as TIFF/EP, was ratified and published. It also supports "non-raw", or "processed", images. TIFF/EP provided a basis for the raw image formats of a number of cameras, but they typically added their own proprietary data. Some cameras have sensors that cannot be described by that version of TIFF/EP. 2004, September: Adobe launched DNG. Its specification states it "is compatible with the TIFF-EP standard". It is a TIFF/EP extension with considerably more specified metadata, brought up-to-date and made fit for purpose. DNG also exploits various other open formats and standards, including Exif metadata, XMP metadata, IPTC metadata, CIE XYZ coordinates, ICC profiles, and JPEG. Although DNG supports more sensor configurations than TIFF/EP (for example, cameras from Fujifilm using Super CCD sensors), it still doesn't support all sensor types as raw images, especially those using the Foveon X3 sensor or similar, hence especially Sigma cameras. 2006: TIFF/EP began its 5-year revision cycle. 2006 to 2007: Adobe offered the DNG specification to ISO to be part of ISO's revised TIFF/EP standard. 2008, September & October: Minutes of ISO/TC 130/WG2 — Prepress Data Exchange, 37th Meeting: "WG 18 is revising the two-part standard (ISO 12234), which addresses digital camera removable memory. The revision of... Part 2 will add DNG into TIFF/EP." A progress report from ISO about the revision of TIFF/EP stated that the revision "...currently includes two "interoperability-profiles," "IP 1" for processed image data, using ".TIF" extension, and "IP 2" for "raw" image data, ".DNG" extension". 2009, September: Minutes of ISO/TC 130/WG2 — Prepress Data Exchange, 39th Meeting: the revision of TIFF/EP "is comprehensive to support many different use cases, including backward compatibility with current TIFF readers and support of Adobe DNG... Profile 2 (proposed extension .dng, if Adobe is in agreement) is intended for camera raw images, including un-demosaiced images... This format will be similar to DNG 1.3, which serves as the starting point for development." See also Comparison of image viewers DNxHD codec dcraw LibRaw References Digital photography Graphics file formats
70076585
https://en.wikipedia.org/wiki/High%20integrity%20software
High integrity software
High-integrity software is software whose failure may cause serious damage with possible "life-threatening consequences." “Integrity is important as it demonstrates the safety, security, and maintainability of your code.” Examples of high-integrity software are nuclear reactor control, avionics software, and process control software. A number of standards are applicable to high-integrity software, including: DO-178C, Software Considerations in Airborne Systems and Equipment Certification CENELEC EN 50128, Railway applications - Communication, signalling and processing systems - Software for railway control and protection systems IEC 61508, Functional Safety of Electrical/Electronic/Programmable Electronic Safety-related Systems (E/E/PE, or E/E/PES) See also High availability software Formal methods Software of unknown pedigree References External links Software by type
50368270
https://en.wikipedia.org/wiki/Troy%20Trojans%20men%27s%20golf
Troy Trojans men's golf
The Troy Trojans men's golf teams represent Troy University located in Troy, Alabama, and compete in National Collegiate Athletic Association (NCAA) Division I and the Sun Belt Conference. The Trojans play their home matches at the Troy Country Club, and use the newly built/renovated Trojan Oaks Golf Practice Facility for practicing. History The Troy men's golf team began playing golf 1951, competing primarily in the Alabama Collegiate Conference. In 1968 and 1969, then player and future head coach Mike Griffin who was the #1 seed player for Troy State, won the Alabama Collegiate Conference individual championship two years in a row. Under head coach Mike Griffin, the Trojan men were one of the most dominant golf teams in the NCAA, winning two Division II national championships and also finishing as a runner-up. In total, the Trojan men have won three NCAA Division II Men's Golf Championships in 1976, 1977, and 1984. They also finished as national runner-up in 1978, 1983, and 1992. From 1975 to 1993, the Trojan men's golf team made an unprecedented 19-straight appearances in the NCAA Division II Golf Championships. During Griffin's coaching tenure, player Paul Perini went on to win the NCAA Division II Individual Championship in 1980. Griffin was inducted into the Golf Coaches Association of America Hall of Fame in 2000 for his accomplishments. In 2002, Matt Terry took over the head coaching position at Troy. He has helped lead Troy from their transition from Division II to Division I, and has had remarkable success doing so, winning eight conference championships since the team joined Division I. Terry has also led Troy to multiple national rankings, and five NCAA Regional appearances. In 2017, player Cam Norman qualified to receive an individual at-large bid to the NCAA Golf Championship held at Rich Harvest Farms. He finished tied for 23rd place, which is the highest a Trojan player has ever finished in individual competition at the NCAA Golf Championship since Troy joined Division I in 1994. National Championships NCAA Division II – national champions: 1976, 1977, 1984 NCAA Division II – national runner-up: 1978, 1983, 1992 NCAA Division II – individual national champions: 1980: Paul Perini Conference championships Division II 1975 Gulf South Conference 1976 Gulf South Conference 1980 Gulf South Conference 1981 Gulf South Conference 1982 Gulf South Conference 1983 Gulf South Conference 1984 Gulf South Conference 1986 Gulf South Conference 1987 Gulf South Conference 1988 Gulf South Conference 1990 Gulf South Conference Division I 1994 East Coast Conference 1995 Mid-Continent Conference 1996 Mid-Continent Conference 1997 Mid-Continent Conference 2016 Sun Belt Conference NCAA appearances Team results Individual results Award winners Elite 89 Award Tolver Dozier – 2015 All-Americans The Troy men's golf team has had 38 players named All-Americans, with 10 of them being named 1st Team All-American. Trojans on professional tours Ben Bates Josh Broadaway Ricky Beck Tolver Dozier Zach Portemont Jake Tucker Trojan Oaks Golf Complex The Trojans golf team's home practice course is the Trojan Oaks Practice Course, located in Troy, Alabama. The facility, which underwent a $1.5 million renovation in 2013, used 40 acres of the original Trojan Oaks Golf Course and created a 9-hole, par-34 practice course plus state-of-the-art putting and chipping greens, a wedge practice area, a full driving range, and a new golf clubhouse. The courses hitting bays feature FlightScope Technology for swing analysis, Sam PuttLab for putting analysis, and BodiTrak monitors to measure the body weight shifting when players swing. It is the only course of its kind on the Sun Belt Conference. References
5707537
https://en.wikipedia.org/wiki/President%27s%20Council%20of%20Advisors%20on%20Science%20and%20Technology
President's Council of Advisors on Science and Technology
The President's Council of Advisors on Science and Technology (PCAST) is a council, chartered (or re-chartered) in each administration with a broad mandate to advise the President of the United States on science and technology. The current PCAST was established by on September 30, 2001, by President George W. Bush, was re-chartered by President Obama's April 21, 2010, , and was most recently re-chartered by President Trump's October 22, 2019, . History The council follows a tradition of presidential advisory panels focused on science and technology that dates back to President Franklin D. Roosevelt's Science Advisory Board, continued by President Harry Truman. Renamed the President's Science Advisory Committee (PSAC) by Dwight Eisenhower, it was disbanded by President Richard Nixon. Reagan science advisor Jay Keyworth re-established a smaller "White House Science Council" It reported, however, to him, not directly to the president. Renamed PCAST, and reporting directly to the president, a new council was chartered by President George H. W. Bush in 1990, enabling the president to receive advice directly from the private and academic sectors on technology, scientific research priorities, and mathematics and science education. Mission The President's Council of Advisors on Science and Technology mission is to provide advice to the president and the Executive Office of the President. PCAST makes policy recommendations in areas such as understanding of science, technology, and innovation. PCAST is administered by the Office of Science and Technology Policy (OSTP). Recent PCAST reports have addressed antibiotic resistance, education technology (with a focus on MOOCs), cybersecurity, climate change, networking and information technology, and agricultural preparedness, among many others. Members and structure PCAST has been enlarged since its inception and currently consists of 18 members plus the director of the Office of Science and Technology Policy, who serves as the council's co-chair. The council members, distinguished individuals appointed by the president, are drawn from industry, education, research institutions, and other NGOs. The council is administered by an executive director. On October 22, 2019, after a record 33 months since President Obama's PCAST held its final meeting, the Trump administration issued an executive order reestablishing the PCAST, appointing its first seven members: Catherine Bessant, the chief operations and technology officer at Bank of America Shannon Blunt, a professor of electrical engineering and computer science at the University of Kansas Dario Gil, an electrical engineer and computer scientist, as well as the director IBM Research Robert Iger, CEO of the Walt Disney Company Dorota Grejner-Brzezinska, a professor of engineering at Ohio State University, as well as associate dean for research Sharon Hrynkow, chief scientific officer at Cylo Therapeutics, Inc., a biotechnology company that focuses on research around rare diseases Herbert Fisk Johnson III, the CEO of S. C. Johnson & Son Abraham (Avi) Loeb, a professor of science at Harvard University, director of the Institute for Theory and Computation and the Black Hole Initiative, and chair of the Board on Physics and Astronomy of the National Academies Theresa Mayer, executive vice president for research and partnerships and professor at Purdue University Daniela Rus, a professor of electrical engineering at MIT, director of the Computer Science and Artificial Intelligence Laboratory A. N. Sreeram, a senior vice president at the Dow Chemical Company with a doctorate in materials science and engineering from MIT Hussein Tawbi, associate professor at the Department of Melanoma Medical Oncology, University of Texas MD Anderson Cancer Center, Houston Shane Wall, the chief technology officer for Hewlett-Packard and director of HP Labs K. Birgitta Whaley, a chemistry professor at the University of California, Berkeley and a scientist at Lawrence Berkeley National Laboratory The council is chaired by Office of Science and Technology Policy Director Kelvin Droegemeier. PCAST membership under President Obama The PCAST under President Obama was co-chaired by John P. Holdren and Eric Lander. The outgoing membership included: John P. Holdren serves as one of two co-chairs of PCAST in addition to his duties as the director of the Office of Science and Technology Policy in the Executive Office of the President and assistant to the president for science and technology. Previously he was a professor of environmental policy and director of the Program on Science, Technology, and Public Policy at Harvard University’s Kennedy School of Government. He also served concurrently as professor of environmental science and policy in Harvard’s Department of Earth and Planetary Sciences and as director of the independent, nonprofit Woods Hole Research Center. He is a member of the National Academy of Sciences, the National Academy of Engineering, and the American Academy of Arts and Sciences, as well as a former president of the American Association for the Advancement of Science and recipient of the MacArthur Foundation Prize Fellowship. Eric Lander serves as one of two co-chairs of PCAST as well as the director of the Broad Institute of MIT and Harvard. He is a professor of biology at MIT and professor of systems biology at Harvard Medical School, and is a member of the Whitehead Institute for Biomedical Research. He was one of the principal leaders of the Human Genome Project, recipient of the MacArthur Foundation Prize Fellowship and is a member of both the National Academy of Sciences and Institute of Medicine. William H. Press is one of the two vice-chairs, and is professor of computer sciences at the University of Texas at Austin, has wide-ranging expertise in computer science, astrophysics, and international security. A member of the U.S. National Academy of Sciences, he previously served as Deputy Laboratory Director for Science and Technology at the Los Alamos National Laboratory from 1998 to 2004. He is a professor of astronomy and physics at Harvard University and a former member of the Harvard–Smithsonian Center for Astrophysics (1982–1998). Maxine Savitz is one of the two vice chairs, and is a retired general manager of Technology Partnerships at Honeywell and has more than 30 years of experience managing research, development and implementation programs for the public and private sectors, including in the aerospace, transportation, and industrial sectors. From 1979 to 1983 she served as deputy assistant secretary for conservation in the U.S. Department of Energy. She currently serves as vice-president of the National Academy of Engineering. Rosina Bierbaum, a widely recognized expert in climate-change science and ecology, is dean of the School of Natural Resources and Environment at the University of Michigan. Her PhD is in evolutionary biology and ecology. She served as associate director for environment in OSTP in the Clinton administration, as well as acting director of OSTP in 2000–2001. She is a member of the American Academy of Arts and Sciences. Christine Cassel is president and CEO of the American Board of Internal Medicine and previously served as dean of the School of Medicine and vice president for medical affairs at Oregon Health & Science University. A member of the U.S. Institute of Medicine, she is a leading expert in geriatric medicine and quality of care. Christopher Chyba is professor of astrophysical sciences and international affairs at Princeton University and a member of the Committee on International Security and Arms Control of the National Academy of Sciences. His scientific work focuses on solar system exploration and his security-related research emphasizes nuclear and biological weapons policy, proliferation, and terrorism. He served on the White House staff from 1993 to 1995 at the National Security Council and the Office of Science and Technology Policy and was awarded a MacArthur Prize Fellowship (2001) for his work in both planetary science and international security. Sylvester James Gates, Jr., is the John S. Toll Professor of Physics and director of the Center for String and Particle Theory at the University of Maryland, College Park. He is the first African American to hold an endowed chair in physics at a major research university. He has served as a consultant to the National Science Foundation, the U.S. Departments of Energy and Defense, and the Educational Testing Service, and held appointments at MIT, Harvard, California Institute of Technology, and Howard University. Mark Gorenberg. is a managing director of Hummer Winblad Venture Partners, which he joined in 1990 when the firm began investing its first fund. Previously, he was with Sun Microsystems, where he managed emerging new media areas and was a member of the original SPARCstation team. Susan L. Graham is the Pehong Chen Distinguished Professor of Electrical Engineering and Computer Science Emerita at the University of California, Berkeley. She has won the Harvard Medal, the IEEE John von Neumann Medal, the Berkeley Citation, and the ACM/IEEE Ken Kennedy Award. She was named a University of California, Berkeley Fellow in 2011. She was a member of the President’s Information Technology Advisory Committee (PITAC) from 1997 to 2003. She served as the Chief Computer Scientist for the National Partnership for Advanced Computational Infrastructure (NPACI) from 1997 to 2005. She currently chairs the Computing Research Association’s Computing Community Consortium. Graham is a member of the National Academy of Engineering and the American Academy of Arts and Sciences, and she is a Fellow of the American Association for the Advancement of Science, the Association for Computing Machinery (ACM), and the Institute of Electrical and Electronics Engineers (IEEE). J. Michael McQuade is senior vice president for science & technology at United Technologies Corporation. Prior to joining UTC in 2006, he served as vice president of 3M’s Medical Division, and before that he was president of Eastman Kodak’s Health Imaging Business. He is a member of the board of trustees for Carnegie Mellon University, the board of directors of Project HOPE, and the board of trustees for Miss Porter’s School. He serves on advisory and visiting boards for a number of university science and engineering schools. He currently serves as a member of the Secretary of Energy Advisory Board. Chad Mirkin is the founding director of the International Institute for Nanotechnology, the George B. Rathmann Professor of Chemistry, professor of chemical and biological engineering, professor of biomedical engineering, professor of materials science & engineering, and professor of medicine at Northwestern University. He is a chemist and a world-renowned nanoscience expert, who is known for his development of nanoparticle-based biodetection schemes, the invention of Dip-Pen Nanolithography, and contributions to supramolecular chemistry. He is one of only fifteen scientists, engineers and medical doctors, and the only chemist to be elected into all three branches of the National Academies, and he has been recognized for his accomplishments with over 90 national and international awards, including the $500,000 Lemelson-MIT Prize, the Linus Pauling Medal, and the Feynman Prize in Nanotechnology. Mario J. Molina is a professor of chemistry and biochemistry at the University of California, San Diego, and the Center for Atmospheric Sciences at the Scripps Institution of Oceanography, as well as director of the Mario Molina Center for Energy and Environment in Mexico City. He received the Nobel Prize in Chemistry in 1995 for his role in elucidating the threat to the Earth's ozone layer of chlorofluorocarbon gases. The only Mexican-born Nobel laureate in science, he served on PCAST for both Clinton terms. He is a member of both the National Academy of Sciences and the Institute of Medicine. Craig Mundie is chief research and strategy officer at Microsoft. He has 39 years of experience in the computer industry, beginning as a developer of operating systems. He co-founded and served as CEO of Alliant Computer Systems. Barbara A. Schaal is professor of biology at Washington University in St. Louis. She is a renowned plant geneticist who has used molecular genetics to understand the evolution and ecology of plants, ranging from the U.S. Midwest to the tropics. She serves as vice president of the National Academy of Sciences, the first woman ever elected to that role. Eric Schmidt is the executive chairman of Google and a former member of the board of directors of Apple Inc. Before joining Google, he served as chief technology officer for Sun Microsystems and later as CEO of Novell Inc. Daniel P. Schrag is the Sturgis Hooper Professor of Geology in the Department of Earth and Planetary Sciences at Harvard University and professor of environmental science and engineering in the School of Engineering and Applied Sciences. He is also director of the Harvard-wide Center for Environment. He was trained as a marine geochemist and has employed a variety of methods to study the carbon cycle and climate over a wide range of Earth’s history. Awarded a MacArthur Prize Fellowship in 2000, he has recently been working on technological approaches to mitigating future climate change. Ed Penhoet is a director of Alta Partners. He serves on the board of directors for ChemoCentryx, Immune Design, Metabolex, and Scynexis. He was a co-founder of Chiron and served as the company’s president and chief executive officer from 1981 until 1998. He was also a member of the Independent Citizens Oversight Committee for the California Institute for Regenerative Medicine (CIRM). From 2004 to 2008 he served as the president of the Gordon and Betty Moore Foundation, where he is currently serving on the board. Penhoet was a faculty member of the biochemistry department of the University of California, Berkeley. From July 1998 to July 2002, he served as dean of the School of Public Health at the University of California, Berkeley. He is a member of the US Institute of Medicine and the American Academy of Arts and Sciences. He serves on the board of Children’s Hospital & Research Center Oakland. See also Office of Science and Technology Policy National Science and Technology Council Technology policy References External links President's Council of Advisors on Science and Technology 2001 establishments in the United States Science and technology in the United States United States national commissions
23780234
https://en.wikipedia.org/wiki/LinSig
LinSig
LinSig is a software tool by JCT Consultancy which allows traffic engineers to model traffic signals and their effect on traffic capacities and queuing. As well as modelling the effects of traffic signals LinSig also optimises signal timings to reduce delay or increase capacity at a junction or group of interlinked junctions. History LinSig was developed in the UK by Brian Simmonite in the 1980s who went on to set up JCT Consultancy. The software has been updated a number of times during its life. DOS Versions LinSig was first written in the mid-1980s as a DOS-based tool for analysing and optimising single traffic signal junctions. The use of the software spread mainly by word of mouth within the traffic signals profession becoming widely used by signal engineers. The LinSig traffic modelling was based on standard 'Y' value hand calculations which were widely used at the time for manually calculating traffic signal capacities and delays. This allowed the results of the software to be easily checked against existing manual calculations. LinSig for Windows Version 1 - Released March 2000 LinSig for Windows Version 1 retained the same modelling capabilities as the earlier DOS versions but provided a Windows interface. This version also introduced graphical diagrams into the software showing information such as the junction's lane layout and the phasing and staging arrangements. LinSig Version 2 - Released April 2006 LinSig Version 2 was the first version of LinSig to allow the modelling of more than one junction. A new network traffic model was introduced which allows the coordination of closely spaced traffic signal junctions to be modelled. LinSig Version 3 - Released June 2009 LinSig 3 changed from a link-based model to a lane-based model. It also allowed the user to model larger networks and provides new network modelling tools such as delay-based traffic assignment and entropy-based trip matrix estimation. Using this software it is possible to model signalised roundabouts. Modelling basis LinSig uses a development of the Cyclic Flow Profile traffic model which has been used extensively in the UK for many years. The Cyclic Flow Profile Model simulates the interaction between closely space junctions allowing coordinated signal timings to be modelled. Typical Input data LinSig input data includes observed traffic flows, traffic signal controller phases and stages, time between green signals, saturation flows and lane lengths. Pedestrian Modelling Pedestrian delays at signal junctions can be modelled for individual or linked crossings. Signal optimisation LinSig optimises signal timings either to maximise traffic capacity or minimise delay at a junction. Network modelling LinSig can model and optimise networks of several junctions as well as individual junctions. It is designed to model small groups of junctions in detail rather than larger networks covering entire towns. Delay Based Assignment LinSig uses a traditional user equilibrium assignment to assign traffic to routes through the network based on delays calculated by the network model. Matrix Estimation LinSig provides a matrix estimation facility to generate a network wide trip matrix from junction turning counts. This uses a combination of traditional entropy-based estimation methods together with customisations targeted at estimating matrices in smaller networks. Alternative solutions The main competitor within the UK marketplace is TRANSYT. See also Traffic light control and coordination References External links JCT Consultancy Transyt Road traffic management Road transport in the United Kingdom Traffic signals Traffic simulation
22522498
https://en.wikipedia.org/wiki/Open%20Language%20Tools
Open Language Tools
Open Language Tools is a Java project released by Sun Microsystems under the terms of Sun’s CDDL (a GPL-incompatible free software license). Open Language Tools are intended for people who are involved in translation of software and documentation into different natural languages (localisation engineers, translators, etc.). They are based around common localisation industry standard file formats such as XLIFF and TMX. Open Language Tools consist of the XLIFF Filters designed to convert different source file formats to XLIFF and the XLIFF Translation Editor which is designed to read and edit XLIFF files. They are written in Java and run on Windows, Mac OS, or Linux as long as Java J2RE (at least 1.4.2) is installed. Workflow XLIFF Filters This is an application designed to convert different source file formats to an XLIFF format. It is currently based around the XLIFF 1.0 specification. The conversion is simple. Launch the filters, drag and drop a source file on the application, and the file will be converted to a .xlz file in the same directory as the source file. What basically happens is that the filter: reads and parses the source file, separates the translatable portions of the text from the non-translatable portions by breaking the file into blocks further segments the translatable blocks into smaller sections, typically sentences, and writes the XLIFF file (.xlf) and the skeleton file (.skl) which are further bundled within a zip archive with the .xlz extension. XLIFF Translation Editor This is an application which is made for translating the contents of XLIFF files. As the XLIFF format itself is quite complex and not very user friendly (especially to people not familiar with XML), XLIFF Translation Editor was developed to make the translation and editing much more feasible. Upon opening it shows the user two major panes with the source and target language. The text is segmented and marked – whether it is translated or untranslated, whether it has any 100% matches or fuzzy translations suggested from other applications that may have processed the XLIFF file in the past. The source and target segments are shown in order to give the translator the full context and they are directly editable. Initially, the translator creates a Project into which either an .xlf or .xlz files are inserted. The central part of the editor is the “Mini-TM”, a simple file-based database with all source and target segments from the current project, which is being automatically updated. As soon as the translator starts translating a segment which has already been translated, the target segment is automatically inserted based on the Mini-TM. Similarly, the Mini-TM suggests fuzzy matches. Segments can be marked as approved, reviewed, unreviewed, problematic, Mini-TM translations or user translations. There is also a mechanism to display notes from the source documents (typically from formats like .po files or from other translators). There is a back-conversion tool integrated into the editor which converts the XLIFF files to their original format. For each XLIFF file, it is possible to create a translation memory (TMX) file, which can be later used to update the central translation database. Supported files Documentation file types HTML DocBook SGML JSP XML (generic - needs a configuration file for each XML type) OpenOffice.org: sxw, sxc, sxi Open Document Format: odw, odc, odi Plain text Software file types PO (gettext) Msg/tmsg (catgets) Java .properties Java ResourceBundle Mozilla .DTD resource files See also Translation memory Computer-assisted translation Office Open XML software OpenDocument software OmegaT References External links Open Language Tools - Official Open Language Tools website Open Language Tools download – Download Open Language Tools Free Language Tools – Free Language tools of Websites User's Manual to XLIFF Translation Editor XLIFF Translation Editor demo Help to improve Open Language Tools Software-localization tools Free software programmed in Java (programming language) Computer-assisted translation software programmed in Java
20378555
https://en.wikipedia.org/wiki/Packet%20boat
Packet boat
Packet boats were medium-sized boats designed for domestic mail, passenger, and freight transportation in European countries and in North American rivers and canals, some of them steam driven. They were used extensively during the 18th and 19th centuries and featured regularly scheduled service. When such ships were put into use in the 18th century on the Atlantic Ocean between Great Britain and its colonies, the services were called the packet trade. Steam driven packets were used extensively in the United States in the 19th century on the Mississippi and Missouri rivers, supplying and bringing personnel to forts and trading posts. History Packet craft were used extensively in European coastal mail services since the 17th century, and gradually added cramped passenger accommodation. Passenger accommodations were minimal: transportation, "firing" (i.e. a place to cook), drinking water (often tasting of indigo or tobacco, which the water casks had previously held), and a place to sleep. Later, scheduled services were offered, but the time journeys took depended much on the weather. In 1724, Daniel Defoe wrote about them in his novel Roxana: The Fortunate Mistress. In the mid-18th century England, the King maintained a weekly packet service with the continent and Ireland using 15 packet vessels. Their importance is evident from the fact that the first craft built in the colony of New South Wales (in 1789) was the Rose Hill Packet. Over the two centuries of the sailing packet craft development, they came in various rig configurations which included: schooners, schooners-brigs, sloops, cutters, brigs, brigantines, luggers, feluccas, galleys, xebecs, barques and their ultimate development in the clipper ships. Earlier they were also known as dispatch boats, but the service was also provided by privateers during time of war, and on occasion chartered private yachts. News of "record passages" was eagerly awaited by the public, and the craft's captain and crew were often celebrated in the press. Behind this search for sailing faster than the wind however lay the foundations for a development in naval architecture and its science which would serve until the appearance of the steam vessels. In 1863, during the Civil War, the packet boat Marshall carried the body of Confederate General Stonewall Jackson from Lynchburg to his home in Lexington, Virginia for burial. Canal packet boats The American canal packet boats were typically narrow (about ) to accommodate canals, but might be long. When the Erie Canal opened in New York state in 1825 along the Mohawk River, demand quickly rose for travelers to be accommodated. Canal packet boats included cabin space for up to 60 passengers. Unlike European and American sailing vessels, that sought to attain greater speed under sail, the canal packet boats were drawn through the Erie Canal by teams of two or three horses or mules. Compared to overland travel, the boats cut journey time in half and were much more comfortable. Travelers could get from New York City to Buffalo in ten days, with a combination of sailing and packet boats. Some passengers took the boats to see both the Erie Canal and the natural landscapes. Significantly, thousands of others used packet boats to emigrate to Ohio and other parts of the Midwest. These boats were also instrumental in the settling of and travel within Upstate New York through the branch canals such as the Chenango Canal. Packet boats were also popular along the James River and Kanawha Canal in Virginia, allowing travel beyond the falls upriver. Atlantic packet ships Packets were the predecessors of the twentieth-century ocean liner and were the first to sail between American and European ports on regular schedules. The first company, the Black Ball Line (later the "Old Line") began operating 1 January 1818, offering a monthly service between New York and Liverpool with four ships. In 1821, Byrnes, Grimble & Co. inaugurated the Red Star Line of Liverpool Packets, with the four ships Panther, Hercules, Manhattan and Meteor. In 1822, Messrs Fish, Grinnell, & Co. began the Swallowtail Line, known as the "Fourth Line of Packets for New York," their first ships being the Silas Richards, Napoleon, George, and York, which soon moved to bi-weekly service. By 1825, vessels were advertised as leaving New York on the 8th and leaving Liverpool on the 24th of every month. Their actual schedules eventually varied, sometimes wildly, due to weather and other conditions. Mail steamer Mail steamers were steamships which carried the mail across waterways, such as across an ocean or between islands, primarily during the 19th century and early 20th century, when the cost of sending a letter was declining to the point an ordinary person could afford the cost of sending a letter across great distances. In addition to carrying mail, most mail steamers carried passengers or cargo since the revenue from the mail service, if any, was insufficient by itself to pay for the cost of its travel. However, the advantage for a steamship carrying mail was that its arrival would be advertised in advance in the newspapers, thus giving it "free advertising" as a travel option for passengers or cargo. In most cases, mail carried by mail steamers was delivered to the post office to which it was addressed. In some cases, the incoming mail would be advertised in the local newspaper for pickup at the post office or at the steamship's office for a fee, if not already fee-paid. Occasionally, because of political instability when a post office could not provide normal services, incoming mail from a mail steamer would be delivered to a local delivery service, which would deliver the mail and charge the addressee an extra fee for the service. When this occurred, the local delivery service would place its own local service stamp or mark on the envelope when the extra fee was paid. Universal Postal Union regulations Mail carried by these steamers – sometimes known as paquebot mail – was subject to various regulations by the governments involved as well as the Universal Postal Union's (UPU) regulations stated at the UPU Vienna Conference of 1891. Aircraft namesake The C-82 Packet twin-engined, twin-boom cargo aircraft designed and built by Fairchild Aircraft was named as a tribute to the packet boat. It was used by the United States Army Air Forces and the successor United States Air Force following World War II. Commerce and journalism Receiving information as quickly as possible—whether regarding particulars about trade, foreign markets, decision-making, professional partnerships, business documents, legal contracts, personal letters and political, government and military news—was of urgent importance to 19th century commerce. Industry and business made special arrangements to beat their competitors so that sailing ships, especially packet ships involved in the packet trade, emerged as the central information superhighway of the era, and for the development of journalism as well. For instance, in late January 1840, the American Packet ship Patrick Henry arrived ahead of schedule and beat the competition to deliver the news from the continent for eager American readers. The Morning Herald (New York), 1 February, on the front page, reported: "The foreign news given today is highly important. Yesterday afternoon, about half past three, we received it at this office being a full hour before any of the Wall street papers had their's—and by five oclock we issued an Extra, to gratify the immense crowd that surrounded our office. One of our clippers left town at 10 o'clock, and boarded the Patrick Henry outside the bar at about one o'clock." The news was advertised as "Ten Days Later From England—Highly Important" and included articles about war preparations by Russia, Queen Victoria's marriage that month, meetings of Parliament and the French Chamber, and the French King's speech. "By the arrival of the Patrick Henry, Captain Delano, we have received immense files of English papers and periodicals, due to the 25th London, 26th from Liverpool and 23rd from Paris…Neither the Cambridge nor the Independence had arrived out on the 26th of Dec. The Patrick Henry had a fine run of nine days to the long(itude) of 38, where she took, on the 4th inst, strong westerly gales, which prevailed since that time without change." Improvements in the speed of that communication was crucial for many commercial, financial and shipping business activities—speedier information made capital move faster, directly affecting world trade. In 1840, the Patrick Henry was among twenty sailing packet ships on the New York–Liverpool run, and notably among the speediest. The short round trips, however, did not depend on speed, but rather changes in the schedule. Efficiency may have been improved by tightening schedules, but this may have exacerbated delays and errors of judgment. For westbound sailings, there was a high risk of disaster. Nearly one packet in six was totally lost in service. This means that out of 6,000 crossings, about 22 ended in such wrecks. More than 600 British ships, of all types, were lost each year in between 1833 and 1835 and 1841 and 1842. The loss of lives varied between 1,450 and 1,560. By the time of the maiden voyage of the Patrick Henry, in 1839, packet captains had begun taking more risks against their competitors as steamships were coming into service. Indeed most shipwrecks took place during the period when the competition between sail and steam was hardest. From a mail, business and journalism transmission point of view, the trend was most alarming. Between 1838 and 1847 no less than 21 mail-carrying ships were lost on the North Atlantic route—two each year on average. Two of the ships were Falmouth packets and two were steamers, while 17 were American sailing packets. Eight were on the New York–Liverpool route, two on the Boston–Liverpool route, two on the New York–London route, and five on the New York–Havre route. Six of the ships just disappeared, and were lost with all hands. It is notable that two out of every three wrecks took place in November–February, indicating that the packet captains took too heavy risks, especially during the rough winter sailings. The only precautionary measure to ensure solid business information transmission across the Atlantic was to send duplicates. This was very typical during the shift period. The duplicates also ensured the fastest possible dispatch of information. Most mail—especially eastwards—was still carried by sailing ships during the first decade after the advent of the transatlantic steamship service. Even if the size of the sailing packets grew markedly, their service speed did not follow the trend after the introduction of steamships on the route in the late 1830s. After 1835, there seems to be no signs of speed improvements. Another phenomenon which indicates that the sailing packets were losing their hold on the first class business—mail, fine freight and cabin passengers—was that they no longer cared about the punctuality of the sailing dates as much as they did in the 1830s. If the reliability of a mail ship service is measured by the regularity of sailings and the safety records, the performance of the American sailing packets in the mid-1840s was noticeably below such expectations. Gallery See also Allan Line Royal Mail Steamers Black Ball Line (trans-Atlantic packet) Flagey building in Brussels, nicknamed "Packet Boat" Ocean liner Pony Express Postal history Royal Mail Ship References Service (Mail) Steamer Service Postal Matters Arrival of the Mail! External links Paquebot mail begins at sea, postmarked on land Glossary of Stamp Collecting Terms Alaska Mail Service: the Mail Steamer Elsie By the 1930s a method of signalling the impending arrival of a mail steamer at Aden was still needed Photographs and pictures Woodcut print 1875 photo of Olive, canal freighter Driver and team Packet (river transport) Age of Sail Postal history Steamships
33090200
https://en.wikipedia.org/wiki/RSCS
RSCS
Remote Spooling Communications Subsystem or RSCS is a subsystem ("virtual machine" in VM terminology) of IBM's VM/370 operating system which accepts files transmitted to it from local or remote system and users and transmits them to destination local or remote users and systems. RSCS also transmits commands and messages among users and systems. RSCS is the software that powered the world’s largest network (or network of networks) prior to the Internet and directly influenced both internet development and user acceptance of networking between independently managed organizations. RSCS was developed by Edson Hendricks and T.C. Hartmann. Both as an IBM product and as an IBM internal network, it later became known as VNET. The network interfaces continued to be called the RSCS compatible protocols and were used to interconnect with IBM systems other than VM systems (typically MVS) and non-IBM computers. The history of this program, and its influence on IBM and the IBM user community, is described in contemporaneous accounts and interviews by Melinda Varian. Technical goals and innovations are described by Creasy and by Hendricks and Hartmann in seminal papers. Among academic users, the same software was employed by BITNET and related networks worldwide. Background RSCS arose because people throughout IBM recognized a need to exchange files. Hendricks’s solution was CPREMOTE, which he completed by mid-1969. CPREMOTE was the first example of a “service virtual machine” and was motivated partly by the desire to prove the usefulness of that concept. In 1971, Norman L. Rasmussen, Manager of IBM’s Cambridge Scientific Center (CSC), asked Hendricks to find a way for the CSC machine to communicate with machines at IBM’s other Scientific Centers. CPREMOTE had taught Hendricks so much about how a communications facility would be used and what function was needed in such a facility, that he decided to discard it and begin again with a new design. After additional iterations, based on feedback from real users and contributed suggestions and code from around the company, Hendricks and Tim Hartmann, of the IBM Technology Data Center in Poughkeepsie, NY, produced RSCS, which went into operation within IBM in 1973. The first version of RSCS distributed outside of IBM (1975) was not a complete networking package. It included uncalled subroutines for functions such as store-and-forward that were included in the IBM internal version. The store-and-forward function was added in the VNET PRPQ, first for files, and then for messages and commands. Once those capabilities were added, “the network began to grow like crazy.” Although at first the IBM network depended on people going to their computer room and dialing a phone, it soon began to acquire leased lines. At SHARE XLVI, in February, 1976, Hendricks and Hartmann reported that the network, which was now beginning to be called VNET, spanned the continent and connected 50 systems. By SHARE 52, in March, 1979, they reported that VNET connected 239 systems, in 38 U.S. cities and 10 other countries. “VNET passed 1000 nodes in 1983 and 3000 nodes in 1989. It currently (1990s) connects somewhat more than 4000 nodes, about two-thirds of which are VM systems.” In comparison, by 1981 the ARPANET consisted of 213 host computers. Both ARPANET and VNET continued to grow rapidly. By 1986, IBM’s Think magazine estimated that VNET was saving the company $150,000,000 per year as the result of increased productivity. Other RSCS Protocol Compatible Networks Due to the key role RSCS played in building networks, the line drivers became known as the "RSCS Protocols". The supported protocols were drawn from other programs. The CPREMOTE protocol may have been the very first symmetrical protocol (sometimes called a "balanced" protocol). To expand the RSCS network to include MVS, Hartmann reverse-engineered the HASP Network Job Interface protocol, which enabled the network to grow rapidly. He later added the JES2 Network Job Entry as an RSCS/VNET line driver. BITNET was a cooperative United States university network founded in 1981 by Ira Fuchs at the City University of New York (CUNY) and Greydon Freeman at Yale University which was based on VNET. The first network link was between CUNY and Yale. The BITNET (RSCS) protocols were eventually ported to non-IBM computer systems, and became widely implemented under VAX/VMS in addition to DECnet (The VAX/VMS NJE protocol stack was known as Jnet). At its zenith around 1991, BITNET extended to almost 500 organizations and 3,000 nodes, all educational institutions. It spanned North America (in Canada it was known as NetNorth), Europe (as EARN), India (TIFR) and some Persian Gulf states (as GulfNet). BITNET was also very popular in other parts of the world, especially in South America, where about 200 nodes were implemented and heavily used in the late 1980s and early 1990s. Over time, BITNET was eventually merged into the Internet. Newer versions of RSCS, as well as Jnet and the various UNIX NJE stacks, provided support for TCPNJE line drivers. Since most sites that were on BITNET also had access to the Internet, the BITNET links that were once ran over leased lines and dialup modems were tunneled over the Internet. It was also not uncommon to run NJE over SNA. Technical Issues R. J. Creasy described RSCS as an operating system and considered it an essential component of the VM/370 Time-Sharing System. "The Virtual Machine Facility/370, VM/370 for short, is a convenient name for three different operating systems: the Control Program (CP), the Conversational Monitor System (CMS), and the Remote Spooling and Communications Subsystem (RSCS). Together they form a general purpose tool for the delivery of the computing resources of the IBM System/ 370 machines to a wide variety of people and computers. ...RSCS is the operating system used to provide information transfer among machines linked with communications facilities." Details of the design of RSCS as a virtual machine subsystem are described in the IBM Systems Journal. From a technical point of view, RSCS differed from ARPANET in that it was a point-to-point "store and forward" network, as such it was more like UUCP. Unlike ARPANET, it did not require dedicated Interface Message Processor or continuous network connections. Messages and files were transmitted in their entirety from one server to the next until reaching their destination. In case of a broken network connection RSCS would retain the message and retry transmission when the remote system became available. VNET vs. ARPANET VNET was the first large-scale connectionless network, making it possible for a computer to join the network using dial-up lines, making connection inexpensive while ARPANET required dedicated 50kb lines at first (later raised to 230KB. Most leased lines at the time typically operated at a maximum rate of 9600 baud. VNET employed a vastly simplified routing and path finding approach, later adopted for the Internet. VNET was a true "distributed control" while ARPANET required a "control" center operated at Bolt, Beranek, and Newman in Cambridge, MA. Notes References See also NETDATA Remote job entry Computer networking Computer printing History of the Internet IBM mainframe software VM (operating system)
41457397
https://en.wikipedia.org/wiki/The%20Interview
The Interview
The Interview is a 2014 American political action black comedy film co-produced and directed by Seth Rogen and Evan Goldberg in their second directorial work, following This Is the End (2013). The screenplay was written by Dan Sterling, based on a story he co-wrote with Rogen and Goldberg. The film stars Rogen and James Franco as journalists who set up an interview with North Korean leader Kim Jong-un (Randall Park), and are then recruited by the CIA to assassinate him. The film is inspired by a 2012 Vice documentary. Rogen and Goldberg developed the idea for The Interview in the late 2000s, with Kim Jong-il as the original assassination target. In 2011, after Jong-il's death and Jong-un's succession as the North Korean leader, Rogen and Goldberg redeveloped the script with the focus on Jong-un's character. The Interview was first announced in March 2013, at the beginning of pre-production. Principal photography was in Vancouver from October to December 2013. The film was produced by Columbia Pictures, LStar Capital and Rogen and Goldberg's Point Grey Pictures, and distributed by Sony Pictures Releasing. In June 2014, the North Korean government threatened action against the United States if Sony released the film. As a result, Sony delayed the release from October to December and reportedly re-edited the film to make it more acceptable to North Korea. In November, the computer systems of Sony were hacked by the "Guardians of Peace", a North Korean cybercrime group. The group also threatened terrorist attacks against theaters showing the film. This led to major theater chains opting not to release the film, and Sony instead releasing it for online digital rental and purchase on December 25, 2014, followed by a limited release at selected theaters the next day. In December 2014, South Korean singer Yoon Mi-rae revealed that the film used her song "Pay Day" without permission, and that she was taking legal action. Yoon Mi-rae and her label Feel Ghood Music reached a settlement with Sony Pictures Entertainment on May 13, 2015. The Interview grossed $40 million in digital rentals, making it Sony's most successful digital release and earned an additional $12.3 million worldwide in box office ticket sales on a $44 million budget. It received mixed reviews for its humor and subject matter, although a few critics praised the performances of Rogen, Franco, Park and Diana Bang. Plot Dave Skylark is the host of the talk show Skylark Tonight, where he interviews celebrities (including Eminem and Rob Lowe) about personal topics and gossip. After Skylark and his crew celebrate producer Aaron Rapaport's 1,000th episode, Rapaport is upset by a producer peer who criticizes the show as not being real news. He voices his concern to Skylark, urging change and he agrees. Skylark later discovers North Korean leader Kim Jong-un is a fan of their show, prompting Rapaport to arrange an interview for him. Traveling to the outskirts of Dandong, China, to receive instructions from Sook-yin Park, a North Korean propagandist, Rapaport accepts the interview on behalf of Skylark. Following Rapaport's return, CIA agent Lacey shows up, requesting they assassinate Kim with a transdermal strip of ricin via handshake and replace him with a North Korean resistance movement to prevent Kim from launching a nuclear missile at the West Coast; they reluctantly agree. Skylark carries the strip inside a pack of gum. Upon arrival in Pyongyang, they are greeted by Sook and taken to the presidential palace outside the city, they are introduced to Kim's personal security officers Koh and Yu who are immediately suspicious of the duo; when Koh finds the strip, mistaking it for gum and chewing it. After making a secret and urgent request for help, Lacey airdrops them two more patches via an UAV, but to get it back to their room, Rapaport has to evade a Siberian tiger and hide the container in his rectum before getting caught and searched by security who don't find the container. The next day, Skylark meets Kim and spends the day playing basketball, hanging out, riding in his personal tank and partying with alcohol alongside a group of escorts. Kim explains he is misunderstood as both a dictator and a failed administrator, and they become friends. At dinner, Koh has a seizure from the ricin, accidentally shooting Yu before dying. The next morning, Skylark feels guilty, discarding one of the ricin strips, then thwarts Rapaport's attempt to poison Kim with the second strip. After a dinner mourning the deaths of the bodyguards, Skylark witnesses Kim's true brutal self as he threatens war against South Korea and everyone who tries to undermine him. Skylark goes on a walk, and discovers that a nearby grocery store is fake, realizing Kim has been lying to him. At the same time, during an attempt to seduce Rapaport (who still has the ricin strip), Sook reveals she despises Kim and apologizes for defending the regime. Skylark returns and tries to get Sook's support to assassinate Kim, but she disagrees, suggesting to instead ruin his cult of personality and show the North Korean people the dire state of the country. The trio secretly devise a plan to expose him on air. Rapaport and Sook later arm themselves with guns and have sex in a weapons storage. Before the broadcast begins, Kim gives Skylark a puppy as a symbol of their friendship. During the internationally televised interview with Kim, Skylark addresses increasingly sensitive topics (including the country's food shortage and US imposed economic sanctions) and challenges his need for his father's approval. Meanwhile, Sook and Rapaport take over the control booth, attacking guards trying to cut off the broadcast. Despite his initial resistance, Kim eventually cries uncontrollably, soiling himself after Skylark sings "Firework" by Katy Perry (knowing of Kim's fondness of her music), ruining his reputation. Betrayed and humiliated, Kim goes into a violent rage, shooting Skylark (whose bulletproof vest saves him), and vowing to get his revenge by launching the missile. Skylark regroups with Rapaport and Sook to escape (alongside the puppy) with the help of a security guard. The trio hijack Kim's personal tank to get to their pickup point, running over several guards on the process. Kim chases the group in a helicopter but is shot down by Skylark before he can launch the missile. With the nuclear launch thwarted, Sook guides Skylark and Rapaport to an escape route, explaining that she has to return to Pyongyang to maintain security. The two are later rescued by SEAL Team Six members disguised as North Korean soldiers. Back in the US, Skylark writes a book about his experience in North Korea, Rapaport returns to work as a producer (and maintains contact with Sook via Skype), while North Korea holds its first democratic elections and become denuclearized under Sook's interim leadership. Cast James Franco as Dave Skylark Seth Rogen as Aaron Rapaport Lizzy Caplan as Agent Lacey Randall Park as Kim Jong-un Diana Bang as Sook-yin Park Timothy Simons as Malcolm Reese Alexander as Agent Botwin Anders Holm as Jake Charles Rahi Chun as General Jong Ben Schwartz as Darryl The film also features cameo appearances from Eminem, Rob Lowe, Bill Maher, Seth Meyers, Joseph Gordon-Levitt, Brian Williams and Scott Pelley. Iggy Azalea, Nicki Minaj, Emma Stone, Zac Efron and Guy Fieri appear in the title graphic card for Skylark Tonight. Production Development Seth Rogen and Evan Goldberg developed the idea for The Interview in the late 2000s, joking about what would happen if a journalist was required to assassinate a world leader. Initially, screenwriter Dan Sterling wrote his script involving a fictional dictator from a fictional country, but Rogen, Goldberg and Sony executives asked him to rewrite the script focusing on Kim. The screenplay was then titled Kill Kim Jong Un. Previous iterations of the story revolved around Kim Jong-il, but put the project on hold until Jong-il died and his son Kim Jong-un assumed power in 2011. Development resumed when Rogen and Goldberg realized that Jong-un is closer to their own age, which they felt would be more humorous. To write the story, co-written with Daily Show writer Dan Sterling, they researched meticulously by reading non-fiction books and watching video footage about North Korea. The script was later reviewed by an employee in the State Department. Rogen and Goldberg aimed to make the project more relevant and satirical than their previous films while retaining toilet humor. They were pleased when former NBA star Dennis Rodman visited North Korea, as it reinforced their belief that the premise of the film was realistic. Pre-production In March 2013, it was announced that Rogen and Goldberg would direct a comedy film for Columbia Pictures in which Rogen would star alongside James Franco, with Franco playing a talk-show host and Rogen playing his producer. Rogen and Goldberg were on board to produce along with James Weaver through Point Grey Pictures, while Columbia was said to finance the $30 million budgeted film. Lizzy Caplan joined the film's cast in October 2013. Caplan signed on to play Agent Lacey, a CIA agent who tries to get Franco's character to assassinate the Kim Jong-un. Randall Park and Timothy Simons signed on to co-star later that month. Park plays the North Korean leader Kim Jong-un and Simons the director of the talk show. Park was the first to audition for the role of Kim and got the part immediately. Before filming began, Park gained 15 pounds and shaved his head to resemble Jong-un's crew cut. His role was praised by critics. Although Rogen and Goldberg wrote the character of Kim as "robotic and strict", Park instead played it "sheepish and shy", which they found more humorous. Diana Bang was cast as Sook-yin Park, for which she was well received by critics. Filming Principal photography on the film began in Vancouver, British Columbia, on October 10, 2013, and concluded on December 20, 2013. There are hundreds of visual effects in the film; a crowd scene at the Pyongyang airport, for example, was digitally manipulated with a shot from 22 Jump Street. Pre-release reaction In June 2014, The Guardian reported that the film had "touched a nerve" within the North Korean government, as they are "notoriously paranoid about perceived threats to their safety." The Korean Central News Agency (KCNA), the state news agency of North Korea, reported that their government promised "stern" and "merciless" retaliation if the film was released. KCNA said that the release of a film portraying the assassination of the North Korean leader would not be allowed and it would be considered the "most blatant act of terrorism and war." The next month, North Korea's United Nations ambassador Ja Song-nam condemned the film, describing its production and distribution as "an act of war" and because of Kim's assassination in the film, "the most undisguised sponsoring of terrorism." The Guardian described Song-nam's comments as "perfect publicity for the movie." Later in July, KCNA wrote to U.S. President Barack Obama, asking to have the film pulled. Shortly before the planned release of the film on December 25, 2014, screenwriter Dan Sterling told Creative Screenwriting: "I couldn't believe that the most infamous man in the world knew about my script – but most importantly, I would never want something I wrote to lead to some kind of humanitarian disaster. I would be horrified if anyone got hurt over this." Release Delay and changes In August 2014, Sony delayed the film's release from October 10 to December 25, 2014. Sony made post-production alterations to the film to modify its portrayal of North Korea, including modifying the designs of buttons worn by characters, originally modeled after real North Korean military buttons praising the country's leaders, and cutting a portion of Kim Jong-un's death scene. Sony Pictures Entertainment hack and threats On November 24, 2014, an anonymous group identifying themselves as the "Guardians of Peace" hacked the computer networks of Columbia Pictures's parent company Sony Pictures Entertainment. The hackers leaked internal emails, employee records and several recent and unreleased Sony Pictures films, including Annie, Mr. Turner, Still Alice, and To Write Love on Her Arms. The North Korean government denied involvement in the hack. On December 8, the hackers leaked further materials, including a demand that Sony pull "the movie of terrorism", widely interpreted as referring to The Interview. On December 16, 2014, the hackers threatened to attack the New York premiere of The Interview and any cinema showing the film. Two further messages were released on December 1; one, sent in a private message to Sony executives, said that the hackers would not release further information if Sony never released the film and removed it from the internet. The other, posted to Pastebin, a web application used for text storage which the Guardians of Peace had used for previous messages, stated that Sony had "suffered enough" and could release The Interview, but only if Kim Jong-un's death scene was not "too happy". The message also threatened that if Sony made another film antagonizing North Korea, the hackers "will be here ready to fight". Distribution The Interview was not released in Japan, as live-action comedy films do not often perform well in the market. In the Asia-Pacific region, it was released only in Australia and New Zealand. Rogen predicted that the film would make its way to North Korea, stating that "we were told one of the reasons they're so against the movie is that they're afraid it'll actually get into North Korea. They do have bootlegs and stuff. Maybe the tapes will make their way to North Korea and cause a revolution." Business Insider reported via Free North Korea Radio that there was high demand for bootleg copies of the film in North Korea. The South Korean human rights organizations Fighters for a Free North Korea and Human Rights Foundation, largely made up of North Korean defectors, planned to distribute DVD copies of The Interview via balloon drops. The groups had previously air-dropped offline copies of the Korean Wikipedia into North Korea on a bootable USB memory device. The balloon drop was scrapped after the North Korean government referred to the plan as a de facto "declaration of war." Cancellation of wide theatrical release The premiere was held in Los Angeles on December 11, 2014. The film scheduled a wide release in the UK and Ireland on February 6, 2015. Following the hackers' threats on December 16, Rogen and Franco canceled scheduled publicity appearances and Sony pulled all television advertising. The National Association of Theater Owners said that they would not object to cinema owners delaying the film to ensure the safety of filmgoers. Shortly afterwards, the ArcLight and Carmike cinema chains announced that they would not screen the film. On December 17, Sony canceled the New York City premiere. Later that day, other major theater chains including AMC, Cinemark, Cineplex, Regal, Southern Theatres as well as several independent movie theaters either delayed or canceled screenings of the film, which led to Sony announcing that they were scrapping the wide theatrical release of the film altogether. The chains reportedly came under pressure from shopping malls where many theaters are located, which feared that the terror threat would ruin their holiday sales. They also feared expensive lawsuits in the event of an attack; Cinemark, for instance, contended that it could not have foreseen the 2012 Aurora, Colorado shooting, which took place at one of its multiplexes, a defense that would not hold in the event of an attack at a screening of The Interview. The cancellation also affected other films in which each films' subject matter is North Korea. An Alamo Drafthouse Cinema location in Dallas planned to hold a free screening of Team America: World Police, which satirizes Kim Jong-un's father Kim Jong-il, in place of its previously scheduled screening of The Interview; Paramount Pictures refused to permit the screening. New Regency pulled out of a planned film adaptation of the graphic novel Pyongyang starring Steve Carell; Carell declared it a "sad day for creative expression". Sony received criticism for canceling the wide release. Guardian film critic Peter Bradshaw wrote that it was an "unprecedented defeat on American turf", but that "North Korea will find that their bullying edict will haunt them." In the Capital and Gizmodo suggested the cancellation caused a Streisand effect, whereby the attempt to remove or censor a work has the unintended consequence of publicizing it more widely. In a press conference, U.S. President Barack Obama said that though he was sympathetic to Sony's need to protect employees, he thought Sony had "made a mistake. We cannot have a society in which some dictator in some place can start imposing censorship in the United States. I wish they'd spoken to me first. I would have told them: do not get into the pattern in which you are intimidated." According to Sony Pictures Entertainment CEO Michael Lynton, the cancellation of the wide release was a response to the refusal of cinema chains to screen the film, not the hackers' threats, and that Sony would seek other ways to distribute the film. Sony released a statement saying that the company "is and always has been strongly committed to the First Amendment ... Free expression should never be suppressed by threats and extortion." The movie was not released in Russia. Revised release After the wide release cancellation, Sony considered other ways to release the film, citing pressure from the film industry, theater owners, and the White House. On NBC's Meet the Press on December 21, Sony's legal counsel David Boies noted that the company was still committed to releasing the film. Sony planned a limited release for December 25, 2014, at more than three hundred American independent and arthouse cinemas. Lynton stated that Sony was trying to show the film to the largest audience by securing as many theaters as they could. Sony released The Interview for rental or purchase in the United States through the streaming services Google Play, Xbox Video, and YouTube on December 24, 2014. It was also available for a limited time on SeeTheInterview.com, a website operated by the stealth startup Kernel.com, which Sony previously worked with to market The Fifth Wave. Within hours, The Interview spread to file sharing websites after a security hole allowed people to download rather than stream the film. TorrentFreak estimated that The Interview had been downloaded illegally via torrents at least 1.5 million times in just two days. On December 27, the North Korean National Defence Commission released a statement accusing President Obama of forcing Sony to distribute the film. The film was released on iTunes on December 28. In the first week of January 2015, Sony announced The Interview would receive a wide theatrical release in the United Kingdom and Ireland on February 6, but it would not be distributed digitally in the UK. The film became available for streaming on Netflix on January 24. Home media Sony released the film on Blu-ray Disc and DVD on February 17, 2015. The home release was packaged as the "Freedom Edition", and included 90 minutes of deleted scenes, behind-the-scenes featurettes, a blooper reel, feature commentary with directors Rogen and Goldberg, and a special episode of Naked and Afraid featuring Rogen and Franco. , the film had earned over $6.7 million in sales in the U.S. Reception Box office and online rentals The Interview opened to a limited release in the United States on December 25, 2014, across 331 theaters and earned over $1 million on its opening day. Variety called the opening gross "an impressive launch for a title playing in only about 300 independent theaters in the U.S." It went on to earn over $1.8 million in its opening weekend, and by the end of its run on January 25, 2015, had grossed $6.1 million at the box office. Within four days of its online release on December 24, 2014, The Interview earned over $15 million through online rentals and purchases. It became Sony Pictures' highest-grossing online release, outselling Arbitrage ($14 million), Bachelorette ($8.2 million), and Snowpiercer ($7 million). It was the top-selling Google Play and YouTube film of 2014. By January 20, 2015, the film had earned more than $40 million from online sales and rentals. Sony expected The Interview to break even through video-on-demand sales and saving millions of dollars on marketing. The National Association of Theater Owners contended that Sony would lose at least $30 million due to poor box office performance. Critical response On review aggregation website Rotten Tomatoes, the film holds a 51% approval rating, based on 154 reviews, with an average rating of 5.70/10. The site's consensus reads: "Unfortunately overshadowed by controversy (and under-screened as a result), The Interviews screenplay offers middling laughs bolstered by its two likable leads." On Metacritic, the film has a score of 52 out of 100, based on 33 critics, indicating "mixed or average reviews." IGN's Roth Cornet wrote that "though it's unlikely to stand out as one of the shrewdest political satires of its time, [it] is a clever, unrestrained and—most importantly—sidesplitting parody that pokes fun at both a vapid media and one of the world's most dangerous dictators." Edward Douglas of ComingSoon.net said the film was "hilarious, but it will probably get us nuked." Jordan Hoffman of The Guardian gave the film three out of five stars and wrote that "if this unessential but agreeable movie really triggered an international response, this is life reflecting art in a major way." Scott Foundas of Variety panned the film as "cinematic waterboarding" and "about as funny as a communist food shortage, and just as protracted", but praised the performances of Randall Park and Diana Bang. Mike Hale of The New York Times also praised Park and Bang, but wrote that "after seeing The Interview and the ruckus its mere existence has caused, the only sensible reaction is amazement at the huge disconnect between the innocuousness of the film and the viciousness of the response." Political response In the wake of the Sony Pictures Entertainment hack, leaks revealed e-mails between Sony Pictures Entertainment CEO Michael Lynton and RAND Corporation defense analyst Bruce Bennett from June 2014. Bennett advised against toning down The Interviews graphic Jong-un death scene, in the hope that it would "start some real thinking in South Korea and, I believe, in the North once the DVD leaks into the North". Bennett expressed his view that "the only resolution I can see to the North Korean nuclear and other threats is for the North Korean government to eventually go away", which he felt would be likeliest to occur following an assassination of Kim. Lynton replied that a senior figure in the United States Department of State agreed. Bennett responded that the office of Robert R. King, U.S. Special Envoy for North Korean Human Rights Issues, had determined that the North Korean statements had been "typical North Korean bullying, likely without follow-up". In an interview with CNN, Bennett said Lynton sits on the board of trustees of the RAND Corporation, which had asked Bennett to talk to Lynton and give his opinion on the film. Bennett felt The Interview was "coarse" and "over the top", but that "the depiction of Kim Jong-un was a picture that needed to get into North Korea. There are a lot of people in prison camps in North Korea who need to take advantage of a change of thinking in the north." Bennett felt that if the DVD were smuggled into the country it might have an effect "over time". Bennett contacted the Special Envoy for North Korean Human Rights Issues, a personal friend of his, who "took the standard government approach: we don't tell industry what to do". Jen Psaki, then a spokesperson for the United States Department of State, confirmed that Daniel R. Russel, the U.S. Assistant Secretary of State for East Asian and Pacific Affairs, had spoken to Sony executives; she reiterated that "entertainers are free to make movies of their choosing, and we are not involved in that". Kim Jong-un threatened "merciless" retaliation for his depiction in the film. Seth Rogen responded, "People don't usually wanna kill me for one of my movies until after they've paid 12 bucks for it." Legacy In Greece, in April 2017, the film's opening scene, depicting a young girl reciting a poem with hate speech, was mistakenly broadcast in the news bulletin of Alpha TV and the news program Live News on Epsilon TV, as a real-life provocative event against the United States. In response to the backlash on various online newspapers, Antonis Sroiter and Nikos Evangelatos, the hosts of the said programs, apologized in posts they made on their social accounts. See also Assassinations in fiction List of films set in or about North Korea Team America: World Police, another comedy film satirizing North Korea The Dictator, a comedy film satirizing Middle Eastern dictators Notes References External links 2014 films 2014 controversies 2014 controversies in the United States 2010s buddy comedy films 2010s adventure comedy films 2010s political comedy films 2014 black comedy films 2014 action comedy films 2010s satirical films American films American buddy comedy films American satirical films American action comedy films American adventure comedy films American black comedy films American political comedy films American political satire films Fiction about assassinations Columbia Pictures films Point Grey Pictures films Cultural depictions of Kim Jong-un English-language films Events relating to freedom of expression Korean-language films Films scored by Henry Jackman Films about assassinations Films about the Central Intelligence Agency Films critical of communism Films directed by Evan Goldberg Films directed by Seth Rogen Films produced by Seth Rogen Films with screenplays by Seth Rogen Films set in 2014 Films set in China Films set in Manhattan Films set in Virginia Films set in Nevada Films set in Pyongyang Films set in North Korea Films set in South Korea Films shot in Vancouver Mass media-related controversies in the United States North Korea–United States relations Film controversies Self-censorship Films about journalism Censored films Films produced by Evan Goldberg Films with screenplays by Evan Goldberg
8649479
https://en.wikipedia.org/wiki/File%20binder
File binder
File binders are utility software that allow a user to "bind" multiple files together resulting in a single executable. They are commonly used by hackers to insert other programs such as Trojan horses into otherwise harmless files, making them more difficult to detect. Malware builders (such as keyloggers or stealers) often include a binder by default. A polymorphic packer is a file binder with a polymorphic engine. It thus has the ability to make its payload mutate over time, so it is more difficult to detect and remove. See also Dendroid (malware) MiniPanzer and MegaPanzer – Trojan horses that used file binders for distribution Potentially unwanted program – sometimes have used file binders for distribution References External links MiniPanzer Source Code on SourceForge Hacking (computer security) Types of malware
5879790
https://en.wikipedia.org/wiki/Community%20source
Community source
Community Source is a type of software development used in colleges and universities that builds on the practices of Free Software communities. The software of these collective efforts are distributed via an approved Free Software Foundation licence. Examples include the Sakai Project, Kuali, and Open Source Portfolio. Copyright for the software is often held by an independent foundation (organized as a 501c3 corporation in the United States) modeled on the contributor agreements, licensing, and distribution practices of the Apache Foundation. Description An important distinctive characteristic of community source as opposed to plain open source is that the community includes some organizations or institutions that are committing their resources to the community, in the form of human resources or other financial elements. In this way, the open source project will have both more solid support, rather than purely volunteer efforts as found in other open source communities, and will possibly be shaped by the strategic requirements of the institution committing the resource. Brad Wheeler has noted that Community Source can be understood as a hybrid model of a community like "The Pub between the Cathedral and the Bazaar” where higher education can really solve its [application] software challenges. References below provide more details regarding the use of community source as a practice in higher education. The Community Source Model in Higher Education (Excerpt from Wheeler, 2007 in References) The Community Source Model is a hybrid model that blends elements of directed development, in the classic sense of an organization employing staff and resources to work on a project, and the openness of traditional open-source projects like Apache. The resulting software is available under an Open Source Initiative (OSI) approved license. The code can be examined, changed, redistributed, sold, or incorporated into other products without fee. Anyone can make changes, and subject to quality review, those changes can be incorporated back into an open-source application for the benefit of all. The distinguishing feature of the Community Source Model is that many of the investments of developers' time, design, and project governance come from institutional contributions by colleges, universities, and some commercial firms rather than from individuals. These contributions may be tendered as the first phase of a project, and then additional work may be contributed on an ongoing, voluntary basis by those institutions with a continuing interest in the project. The project often establishes a software framework and baseline functionality, and then the community develops additional features as needed over time. Community Source Model projects generally operate as follows. Several institutions realize they are trying to solve a similar problem—need for a research administration system is a recent example. After some discussions and resulting agreement on project objectives, timelines, and philosophy, the institutions pool their resources under a project board of institutional leaders. The institutions are often agreeing to tender existing staff time to the direction of the project, and as such, this is not a new cash outlay but rather an aggregation of existing staff in a virtual organization. A grant from a foundation may provide cohesion among the investors. Typical recent projects have ranged from $1 to $8 million in funding and from twelve to thirty months in duration. Each investor signs a Corporate Contributor Agreement that grants a copyright license for the software to the project or foundation (modeled on the practice of the Apache Foundation). The project usually operates on a date-driven delivery schedule. This forces difficult decisions in the reality triangle of balancing features, resources, and time, but such a schedule is essential to the growth of community confidence. The project board then establishes the appropriate structure for articulating the system requirements, the technical choices, and a project manager. It is essential that clear roles and responsibilities be established early, and the project participants will benefit from spending some face-to-face time together at the beginning of the project. Experience reveals that some staff members may not work well in distributed, virtual organizations, whereas others find the work to be career-renewing. Early projects had to transition from an investor-based project to a community and a foundation. New projects can take advantage of the foundations' existing infrastructure and how-to knowledge and can begin as a project of a foundation. There is no rulebook for Community Source Model projects for every domain, but there is a growing body of accumulated wisdom on how to coordinate institutional investments and execute a development plan for quality software. Community source software licensing Community source software licensing is when the source code to proprietary software is licensed to members of a defined community, each member of which must explicitly enter an agreement with the code owner in order to be permitted access to source code. By this definition, community source licenses are incompatible with both open source and free software, since in open source and free software anyone may have access to source code without entering into an agreement with anyone else, though they must accept the license. A community source license cannot meet the requirements of the Open Source Definition, whose first provision requires free redistribution of software. A community source license cannot meet the requirements of The Free Software Definition, since freedoms 1 (freedom to study) and 3 (freedom to modify and distribute modified versions to anyone) require open access to source code. A community source license may violate other provisions of the open source and free software definitions. One motivation for community source may be to promote the use of software production and development models similar to those used in open source communities. But because the community is not open, those production and development models must differ in subtle or marked ways from those used in open source. Another motivation may be to ease adoption by customers of complex software from a proprietary supplier, using trusted intermediaries to help. Examples of community source licenses Sun Community Source License RealNetworks Community Source License Microsoft Community License See also Alternative terms for free software List of commercial software with available source code List of commercial video games with available source code References Community-source development appeals in tough times, Darryl K. Taft (March 2, 2010), eWeek The Inevitable Unbundling of Software and Support, Brad Wheeler (February 2004). Campus Technology An Open Mind on Open Source, Karla Hignite (August 2004), Business Officer Magazine Community Source Springs Forth, Anna Jackson (May 2005), Business Officer Magazine A Flexible Financial System, Kathleen McNeely & Mary Wheeler (February, 2006), Business Officer Magazine Open Source 2010: Reflections on 2007, Brad Wheeler (January/February, 2007). EDUCAUSE Review Invest Locally, Christopher Coppola (April 2007), Campus Technology Mitigating the Risks of Big Systems, Brad Wheeler & Joanne DeStefano, (July 2007), Business Officer Magazine Free and open-source software organizations Software licenses
19518793
https://en.wikipedia.org/wiki/Australian%20Computers%20in%20Education%20Conference
Australian Computers in Education Conference
This National Conference is the biennial conference of the Australian Council for Computers in Education (ACCE). The conference opens to anyone who in interested in sharing their digital teaching experiences. The first conference took place in Melbourne, 1983. Between 1983-1996, the conference was held annually across Australia. After 1996, the conference became biennial. From 1994, a series of frameworks were launched in Australia to integrate Information and Communication Technology(ICT) into education. Western Australia's 2001 Competency framework for Teachers identified teachers as an important component in developing computer education. In 2010, Education Minister Julia Gillard, proposed an education agenda to provide Australia a better education system. Besides ACCE, there are many organizations and conferences supporting the development of computer education in Australia. Technology in education consists of two major approaches: Learning with technology and learning from technology. Technology in education learning and traditional classroom learning have different focuses and defining features. There are also four types of computer education:Bring your own device(BYOD), blended learning, online learning, and flipped learning. Purpose The Australian Computers in Education Conference (ACEC) is open to anyone in the education field that wishes to share their digital experiences and is interested in using ICT to promote and enhance classroom learning. The conference aims to oversee the global trends and activities and compare them to the work accomplished by the Australian Computers in Education groups. History The first conference (labelled as ACEC) took place in Melbourne in 1983. Since that time it has continued to provide teachers around the country a forum to exchange information on classroom activity, make connections with each other, and learn from experts in the field. The conference proceedings provide an insight into the nature of educational computing practice since the early 1980s and the papers are widely cited in the research field literature. The conference was formalized in 1985. A meeting of Victoria Computer Education Group state presidents was held in Melbourne in November. A constitution was formulated, a secretariat was established, and the National Committee for Computers in Education took on a new name: the Australian Council for Computers in Education (ACCE). The meeting also decided to establish a national journal, Australian Educational Computing, under an editorial board. The 5th conference was held in Adelaide in September 1987. It was hosted by the Conference in Education Group of South Australia. The major theme of the conference was Tomorrow's Technology Today. Within the major theme, there were a number of strands being addressed: special education, computer graphics, computer and equity, and staff development and teacher training. The 11th conference was held in Brisbane in July 1993. It was hosted by the Computers in Education Group of New South Wales. 43 papers were presented at the conference, focusing on “research and scholarship in the use of computers in elementary, secondary and higher education level". In 2008, the conference was held in Canberra and the Education Minister Julia Gillard delivered the opening Keynote Address, highlighting the importance of digital technologies in her education reform agenda. In 2012, the conference was held in Perth. The theme of the conference was It's Time. The conference addressed the role of community is teachers’knowledge development and the role of teachers in technological development in education. Teachers were identified as both knowledge providers and facilitators of learning and e-learning process. In 2014, the conference was held in Adelaide. The theme of the conference was Now It's Personal. Papers presented at the conference focused on the positive and negative impacts of ICT projects on student attitudes and capabilities, students’conceptual understanding of digital devices, and the importance of harmonizing pedagogy with technology. Conference locations 2016: Brisbane, theme: IF 2014: Adelaide, theme: Now ITs Personal. Venue: Adelaide Convention Centre 2012: Perth, theme: ITs Time. Venue: Wesley College, South Perth. 2010: Melbourne, theme: Digital Diversity. Venue: Melbourne Convention and Exhibition Centre 2008: Canberra, theme: "ACT on IcT" Venue: National Convention Centre 2006: Cairns, theme: IT's up here for Thinking. Venue: Cairns Convention Centre 2004: Adelaide 2002: Sandy Bay, Tasmania, theme: "Linking Learners" 2000: Melbourne, theme: Learning Technologies, Teaching and the Future of Schools. Venue: Carlton Crest Hotel. Website: conference website 1998: Adelaide 1996: Canberra 1995: Perth, theme: Learning without limits. Frameworks Australia has six states and two territories. Each of them has a certain degree of independence. Constitutionally, each state has an independent education system and the government has a limited role in coordinating and funding specific computer education projects. ICT and Learning Technologies(LT) in Australia are referred to the use of technologies in teaching and learning activities. Western Australia uses LT to refer to ICT as tools to improve the quality of teaching and learning. In Victoria, LT is referred to as different forms of ICT to support classroom learning. Computers have become commonly used in Australian schools since the 1990s. The 1994 Statement on Technology for Australian schools outlined four strands for ICT in education: Designing, Making and Appraising; Information; Material; and Systems. “The statement also included the place of technology in society, the need of students to experience technology education and the form in which it should appear in the school curriculum.” In 1995, each state of Australia implemented its own strategic plan for using technology to support the library system, school administration and classroom learning. Each state also had their own emphasis in the use of technology in education: Queensland focused on databases, Tasmania focused on desktop publishing and Western Australia focused on tutorial packaging. In 1997, Queensland set up the minimum professional standards for teachers in LT. Teachers were expected to understand the needs of students relative to the use of ICT, select relevant teaching strategies with the use of ICT to cater the needs of students, and create learning experiences where students can use ICT to interpret, analyze and represent knowledge. In August 1999, all Australian states adopted the national goals to transform students into "confident, creative and productive users of new technologies". In the same year, the federal report,Teacher for the 21st Century, recognized the importance of ICT in the professional development of teachers. In Western Australia, the 2001 Competency Framework for Teachers outlined the anticipation for teachers to evaluate LT through the application of personal knowledge, encourage students to select appropriate LT to enhance their learning experiences, and provide opportunities for students to use technology for a variety of purposes. In 2008, the Council of Australian Governments (COAG) released the Melbourne Declaration on Education Goals for Young Australians. The Declaration asked all Australian governments to commit in the pursuit of equal education opportunities and outcomes, relative to the goals of 1999 to transform students into creative and confident users of technology and informed citizens. Education Minister Julia Gillard's 2008 reform agenda highlighted the importance of ICT in education. The agenda aimed to improve the quality of education by recruiting high-performing graduates into teaching and recognizing the high-quality teachers. The vision of the agenda was to create a better education system in Australia. In 2011, the Digital Education Advisory Group (DEAG) was established. The group consists of experts from the government, education, academia and industry field. Their task was to assess the achievements of computer education, analyze what still remained to be achieved, set up new priorities and implement strategies for future developments of digital education. The assessment required examinations of digitcal education policy outcomes and published data of Australian and global initiatives. The National Digital Economy Strategy (NDES) was a strategic framework for Australia to become a world-leading digital economy by 2020. One of the eight goals was to expand online education such that Australian schools, universities and other higher education institutions will have access to online educational services and resources. In 2012, the project called "Digital Education Revolution" (DER) has delivered 911,000 laptops and computers to students in Years 9-12 across Australia for educational purposes for all the subjects at school. ACCE and other supporting organizations and conferences In Australia, there are not many formal national associations to support computer education. The ACCE is one of the formal national organizations on research, learning and innovation. It is the only national organization that is entirely devoted to computer education. ACCE has affiliates in each state. An example is the New South Wales Computer Education Group and Victoria Computer Education Group. The state affiliates take turns to host the ACEC and publish newsletters. Australian Computer Society (ACS) also gives support to computer education and is linked to the International Federation of Information Processing (IFIP). There is a cooperation between organizations and sectors to accelerate the development of computer education. An example is the creation of Asia-Pacific Information Technology in Training and Education (APITITE), consists of ACCE, ACS and Technology for Training and Education (ITTE). Decision has been made by APITITE to combine conferences of the major organizations. The first conference of APITITE was held in June and July in 1994. There are several organizations that support the development of computer education of Australian higher education. Two organizations that are primarily focusing on the promotion of computer education are the Australian Society for Educational Technology (ASET) and the Australian Society for Computers in Learning in Tertiary Education (ASCILITE). These two organizations host annual and biennial conferences which have attracted both local and international delegates. Australasian Computing Education Conference (ACE) was established in 1996 and has provided a forum for academics and experts who are interested in computer education. The conference is primarily for researchers in computer education and educators that are concerned about teaching computing or using computers in their teaching. Technology in education Technology in education has two major approaches. One is learning from technology and one is learning with technology. Learning with technology is known as integrated learning system while learning from technology is known as constructivist environments. Traditional classroom learning is teacher-centred and technology in classroom learning is student-centred. Traditional classroom learning focuses on mass instruction where one size fits all; technology in classroom learning focuses on mass customization to fit individual student and teacher needs. There is insufficient evidence to show that technology in the classroom learning will enable more learning compared to traditional classroom learning and alternative methods. The common forms of technology used in classroom are projectors, desktops, laptops and mobile devices. Artificial Intelligence(AI) is the use of computers to complete the tasks that previously required mankind. AI has already been applied in education in some places and is expected to be applied on a larger scale in the future educational models to fill the gaps in education system. There are three categories of AI software applications in education that are available today: personal tutors, intelligent support for collaborative learning, and intelligent virtual reality. The development of technology has displayed various advantages of integrating technology in education such as increased equity of access, repeatability and transportability. Previous empirical studies have revealed a total of 123 technology integration barriers in education. There are 6 main categories that hinder technology integration: Resources, knowledge and skills, institutions, attitudes and beliefs, assessment, and subject culture. Of the six main barriers, 40% of the studies have identified resources as the primary barrier. There are four types of computer education in classrooms. Bring your own device (BYOD) In the BYOD environment, students bring their own digital device for educational purposes. BYOD allows school to reduce costs of supplying digital devices and allows students to use devices in which they are familiar with. By using their own devices security in social or online world is assured by almost 100%. Blended learning Blended learning is "an education environment where teachers use digital technology in traditional or flipped classrooms on a regular basis". It combines face-to-face learning environment with online learning environment. Blended learning is utilized to eliminate the weakness of fully online learning environment such as isolation. There are a number of variables and aspects involved in determining the structure of a blended learning environment. Decisions about structure are made with respect to the goals and capabilities of the instructor, the students and the learning environment. Online learning Online learning refers to distant education. The majority or all of the course content is delivered by the instructor to students online with no or limited face-to-face interaction and teaching. Flipped learning Flipped learning is a pedagogical model in which direct instruction moves from group learning space to individual learning space to allow instructors to focus on student engagement and active learning. See also C2SV NexCon TechfestNW DEMO conference References External links ACEC 2014 official site ACEC 2012 official site ACEC 2010 official site ACCE official site Education in Australia Technology conferences Recurring events established in 1983 1983 establishments in Australia
67490143
https://en.wikipedia.org/wiki/Romanian%20rural%20systematization%20program
Romanian rural systematization program
The Romanian rural systematization program was a social engineering program undertaken by Nicolae Ceaușescu's Romania primarily at the end of the 1980s. The legal framework for this program was established as early as 1974, but it only began in earnest in March 1988, after the Romanian authorities renounced most favoured nation status and the American human rights scrutiny which came with it. The declared aim of this program was to eliminate the differences between urban and rural, by the means of razing half of Romania's 13,000 villages and moving their residents into hundreds of new "agro-industrial centers" by 2000. The program gained notoriety in Europe, with protests from multiple countries – chiefly Hungary – as well as a Belgian-led initiative to save the Romanian villages by "adopting" them. Within a year, on 18 April 1989, the first batch of 23 new agro-industrial towns was completed. Only one new town was created between 1974 and 1988, as Ceaușescu focused his attention on other projects. Although cut short by the Romanian Revolution in December 1989, at least three more rural settlements in an advanced state of systematization were, ultimately, transformed into towns as well. Background Nicolae Ceaușescu's 1988 idea to raze about half of Romania's 13,000 villages and rebuild others into "agro-industrial centers" was not new. It had been written into law in 1974. At that time, about 3,000 villages were scheduled to die out gradually, while 300–400 more were to be transformed into towns. However, industrial construction assumed priority, overshadowing the rural reconstruction and resettlement program, which was not pursued with any vigor. In the spring of 1988, however, the rural systematization program reemerged as a top priority on Ceaușescu's agenda. The concept was first developed by Nikita Khrushchev, aiming to raise the standard of rural life by amalgamating villages in order to stop the migration of younger people from rural to urban. However, the project was forgotten while Ceaușescu focused on other projects, such as the Centrul Civic and the Danube–Black Sea Canal, but it was relaunched in March 1988. The 1974 law for urban and rural territorial reorganization provided for the development of the countryside by focusing on the more viable villages while the rest would be gradually starved of investment. However, momentum was lost in the late 1970s, and of the 140 new towns promised by 1985, only one – Rovinari – was completed in 1981. No explanation was ever given, but likely Ceaușescu transferred his attention to the aforementioned projects. Details of the plan The villages most likely to be phased out were those with minimal prospects for growth. By the year 2000, 85% of communes were to have piped drinking water and 82% modern sewage. According to a statement by the regime, by the year 2000, Romania expected "to eradicate basic differences between villages and cities and to ensure the harmonious development of all sections of the country". Ceauşescu's declared aim - based on an original idea in the Communist Manifesto - was "to wipe out radically the major differences between towns and villages; to bring the living and working conditions of the working people in the countryside closer to those in the towns". He thought that by gathering people together into apartment buildings so that "the community fully dominates and controls the individual", systematization would produce Romania's "new socialist man". Ceaușescu was determined to revolutionize agriculture by increasing the cultivation area, while also stifling individual initiative and increasing centralization. The peasants were to receive derisory compensation for their demolished homes and then be charged rent for their new blocks, in which there was no accommodation for animals. As Romanian historian Dinu Giurescu put it: "The ultimate goal is the proletarianization of our society. The final step in this process is the loss of the individual house.". It was an all-out effort at social engineering: kitchens and bathrooms were communal space in the government-owned and controlled apartments. The number of villages was to be reduced to 5,000–6,000 (grouped in 2,000 communes), implying that 7,000–8,000 would be destroyed. Workers and intellectuals were to be settled in 3–4 storey buildings, with small blocks of 4 apartments or individual two-storey houses for the farmers. The countryside would be urbanized through 558 new agro-industrial towns. Although aspects of the program were absolutely necessary (improvement of services, diversification and stabilization of the workforce), it allowed little scope for local consultation and its implementation timespan was far too short (hence compulsory resettlement) with no realistic compensation for the required expropriation. The last stop: MFN status Between 3 August 1975 and 3 July 1988, Romania was accorded most favoured nation status from the United States. In 1988, Ceaușescu renounced Romania's MFN status with the United States, just as the latter was about to suspend it over human rights violations. In July 1987, the United States Congress voted to suspend Romania's MFN status. Although the suspension was meant to last at least 6 months, in order to avoid further humiliation, Ceaușescu renounced his country's MFN status. More than 85 oral testimonies and 995 written statements were submitted to support the suspension of Romania's MFN status. On 26 February 1988, in order to save face, Romania announced that it did not need MFN status. The House and Senate votes were rejected as unacceptable "interference in the internal affairs" of Romania. To underline this rejection, the village-bulldozing program was made public in April 1988. Romania's renunciation of MFN status in February 1988 resulted from Ceaușescu's growing irritation with American pressure over Romania's human rights situation, such as Ceaușescu's treatment of his opponents. Ceaușescu's renunciation of MFN made its suspension by the United States Congress meaningless. His action showed that he would not submit to pressure from either side, East or West. Implementation and results Ceaușescu felt fed-up by continuous United States Congressional scrutiny of Romania's human rights record, a scrutiny hindering his long-cherished "grand design". Shortly after his "cocky" gesture on MFN, Ceaușescu announced the most sweeping and ominous plan of his regime up to that point, involving the liquidation of up to 8,000 villages. On 3 March 1988, speaking at an official conference, Ceaușescu announced: by the year 2000, 7,000 - 8,000 of Romania's 13,123 villages would be "modernized", as in transformed into 558 "agro-industrial" centers. The Ilfov Agricultural Sector around Bucharest was chosen by Ceaușescu as a showpiece (to be completed by 1992-1993), as a model for emulation by the rest of the country. The first evictions and demolitions took place in August 1988. Only 2-3 days were given before shops were closed down and bus services were stopped, forcing the inhabitants into the selected villages. Whole communities were moved to blocks in Otopeni and Ghermănești, where as much as 10 families had to share one kitchen and the sewage system had not been completed. In other villages across the country, "ugly" concrete Civic Center buildings began to emerge in the centers of the planned new towns. Around 18 villages had suffered major demolitions by the end of 1989 while 5 others were completely razed. According to the Wall Street Journal: "In the countryside, smashed hamlets and villages are making way for the same prefabricated housing blocks of Orwellian Bucharest.". The systematization program encountered resistance from villagers and local authorities alike. Local revolts against systematization were reported in the villages of Petrova, Monor and Parva. Local officials were threatened, while in other places officials refused to carry out orders. The director of the Miercurea Ciuc County Savings Bank resigned in protest over pressure to designate his native village of Păuleni-Ciuc a street of the nearby town of Frumoasa. The systematization program was terminated on 26 December 1989, the day after the trial and execution of Nicolae and Elena Ceaușescu. Towns created under Systematization The program fell behind schedule, with only 24 new towns declared in 1989 out of the 100 expected by 1990. These 24 agro-industrial towns are listed below: Transylvania Bihor County Valea lui Mihai - Although agriculture remained predominant, small-scale industries were being developed. Around 30 apartment blocks were built in the center. Hunedoara County Aninoasa - A mining center in the upper Jiu Valley. Maramureș County Seini - Main industries: agriculture, animal breeding and fruit growing. Mureș County Iernut Sibiu County Avrig - The Mechanical Works at Mârșa () was the most important industrial plant in the area. Tălmaciu - Several timber factories and a textile plant. Blocks totalling 480 apartments were built. Wallachia Argeș County Colibași - Location of Automobile Dacia, Romania's first car manufacturing plant. Two national research institutes - for automotive engineering and nuclear technology - were also located there. Brăila County Ianca - A model for agro-industrial towns in flat regions. Însurăței - Blocks totalling 330 apartments were built, of which 200 were located in the center. Buzău County Nehoiu - A model for agro-industrial towns in mountainous regions. Pogoanele - Blocks totalling 200 apartments were built in the center. Călărași County Budești Lehliu Gară - Its main industrial plant was a subsidiary of a clothes factory based in Bucharest. Fundulea - Three important agricultural research institutes were based there, most of their staff commuting from Bucharest. Giurgiu County Bolintin-Vale - A satellite of Bucharest. Mihăilești - Most of the former village razed and rebuilt to make way for the Danube–Bucharest Canal. Visited by Ceaușescu multiple times. Blocks totalling 3,500 apartments were built. Gorj County Bumbești-Jiu Rovinari - Already a town as of 9 December 1981. Olt County Piatra Olt Scornicești - The town with the most impressive record of medals, distinctions and titles in Socialist Romania, including that of "Hero of the New Agrarian Revolution". The new town had a factory making automobile spare parts, a clothes factory, a brewery, a dairy plant and a poultry farm. Dobruja Constanța County Basarabi - An inland harbor for the Danube–Black Sea Canal. There were also two large wineries. Negru Vodă - Blocks totalling 341 apartments were built. Ovidiu Moldavia Bacău County Dărmănești Other rural settlements decisively impacted by Systematization which later became towns Bragadiru, Cornetu, Balotești and Otopeni were likewise to become agro-industrial towns. Two hundred dump trucks were required to carry the rubble resulted from the demolition of many private houses in Otopeni, Dimieni and Odăile. In Bragadiru, Măgurele, Otopeni and 30 Decembrie blocks totalling thousands of apartments were built. Otopeni became a town on 28 November 2000. Bragadiru and Măgurele became towns on 29 December 2005. International reactions Hungary In Hungary, the program is called "romániai falurombolás" (lit. "Romanian village destruction"). After May 1988, Transylvanian "atrocity stories" abounded in the Hungarian press. The Romanian regime drew unfavorable world opinion and came under increasing attack from the global press. In June 1988, 50,000 people protested in Budapest because the thousands of villages proposed for destruction by the Romanian Government included 1,500 ethnic Hungarian ones. That demonstration, taking place on 27 June, was the largest organized in Hungary after 1956. The plan was also criticised by leading members of the Hungarian Socialist Workers' Party. In retaliation for the 27 June protest rally held in Budapest, Ceaușescu closed the Hungarian consulate in Cluj. To address these issues, and to comply with Mikhail Gorbachev's request, Károly Grósz met with Ceaușescu in Arad on 28 August 1988. The talks were fruitless, with Ceaușescu unwilling to concede or compromise on any point. He simply used the fact that talks were being held at all as a means to win time and regain some of his lost credibility. The meeting was seen in Hungary as a capitulation of the country's national interests and significantly damaged Grósz's own prestige. Opération Villages Roumains The scale of the potential destruction caused an international outcry to such an extent that it led to the creation of organizations such as the Belgian-based Opération Villages Roumains, which provided for the twinning of threatened Romanian villages with Western communities. Few of the villages were actually destroyed, systematization only really succeeding in "imprinting Romania onto the consciousness of Europe". Opération Villages Roumains was founded on 22 December 1988 in Brussels, being officially launched on 3 February 1989. By the beginning of May 1989, Romanian villages had been adopted by 231 communes in Belgium, 95 in France, and 42 in Switzerland. The Belgian effort was almost exclusively Walloon. In its March 1989 session, the Council of Europe strongly condemned the liquidation of the villages, asking the authorities to cease the campaign. Shortly afterwards, in a speech broadcast by the BBC, the Prince of Wales publicly criticized this policy. References Socialist Republic of Romania Romanian society Urban planning in Romania Nicolae Ceaușescu Social engineering (political science)
57090116
https://en.wikipedia.org/wiki/Customer%20identity%20access%20management
Customer identity access management
Customer (or consumer) identity and access management (CIAM) is a subset of the larger concept of identity access management (IAM) and is focused specifically on managing the identities of customers who need access to corporate websites, web portals and webshops. Instead of managing user accounts in every instance of a software application of a company, the identity is managed in a CIAM component, making reuse of the identity possible. The biggest differentiator between CIAM and regular (internal) IAM is that in CIAM the consumers of the service manage their own accounts and profile data. CIAM functionality Generally speaking a CIAM environment serves the following purposes: Identity as a Service, for managing digital customer identities CRM (Customer Relationship Management), for managing user behaviour Consent Management for managing user consent in reference to Privacy Identity as a Service CIAM is a required component of modern user engagement allowing organizations to recognize unique customers and personalize their engagement based on collected personal preferences. A single CIAM system can control access to multiple applications, using federation protocols to transfer the digital identity and access parameters to the different applications. CIAM solutions are generally designed to scale to handle tens-of-millions of users or more in B2C environments. IAM is common in large organizations to control a wide scope of internal user access points including computer hardware access, file and resource permissions, network access permissions, application access, and human resource needs. In the simplest form, CIAM includes the registration and login processes that allow a customer to sign in and use a company’s application. More advanced systems can provide single sign-on (SSO), account and preference management, data tracking and reporting, multi-factor authentication, and user monitoring and management. CRM The digital identities managed by a CIAM solution are used to give access to different business applications, portals and webshops. Due to the fact that all these transactions are logged, the data can be used for profiling purposes. And transaction data can be correlated to the digital identities of the customers. The data can be seen as a relevant component of CRM systems. Consent Management Because of the nature of CIAM – user logging in, managing profiles, accessing services – CIAM solutions harvest a lot of personal information. Privacy laws, such as the GDPR in the European Union, hold CIAM providers accountable for processing this kind of data, hence the providers have taken steps to restrict the processing of these data by implementing Consent Management services. For every data element users can define whether a provider can process or transfer the personal data. For instance: a user can give or revoke consent to process transaction data for marketing purposes. See also Digital identity Electronic authentication Federated identity Identity assurance Identity management Privacy by design Strong authentication References Identity management
298754
https://en.wikipedia.org/wiki/HP%20Multi-Programming%20Executive
HP Multi-Programming Executive
MPE (Multi-Programming Executive) is a discontinued business-oriented mainframe computer real-time operating system made by Hewlett-Packard. While initially a mini-mainframe, the final high-end systems supported 12 CPUs and over 2000 simultaneous users. Description It runs on the HP 3000 family of computers, which originally used HP custom 16-bit stack architecture CISC CPUs and were later migrated to PA-RISC where the operating system was called MPE XL. In 1983, the original version of MPE was written in a language called SPL (System Programming Language). MPE XL was written primarily in Pascal, with some assembly language and some of the old SPL code. In 1992, the OS name was changed to MPE/iX to indicate Unix interoperability with the addition of POSIX compatibility. The discontinuance of the product line was announced in late 2001, with support from HP terminating at the end of 2010. A number of 3rd party companies still support both the hardware and software. In 2002 HP released the last version MPE/iX 7.5. Commands Among others, MPE/iX supports the following list of common commands and programs. =SHUTDOWN BASIC CHDIR COPY DEBUG ECHO ELSE EXIT FORTRAN HELP IF PASCAL PRINT RENAME SH WHILE See also HP 3000 References External links Allegro Consultants, Inc. Free HP 3000 Software, MPE Software Support Beechglen Development Inc. MPE Software Support HP MPE/iX homepage HP MPE/iX Command reference openMPE Advocates of continued MPE and IMAGE source code access beyond 2010 Discontinued operating systems Multi-Programming Executive Proprietary operating systems 1974 software
68383133
https://en.wikipedia.org/wiki/IBM%20System
IBM System
IBM System - is a common name for IBM products. Digit-named series Hardware IBM Personal System/1 IBM Personal System/note IBM Personal System/2 IBM Personal System/2 note IBM Personal System/2 laptop IBM System/3 IBM System/4 Pi IBM Office System/6 IBM System/7 IBM System/23 IBM System/32 IBM System/34 IBM System/36 IBM System/38 IBM Personal System/55 IBM Personal System/55 note IBM System/88 IBM System/360 IBM System/370 IBM System/390 IBM System/390 Multiprise IBM Advanced System/400 IBM System Cluster 1350 IBM 1800 Data Acquisition and Control System IBM 3850 Mass Storage System IBM 5520 Administrative System IBM RISC System/6000 IBM 7700 Data Acquisition System IBM 8100 Information System IBM System 9000 IBM Enterprise System/9000 Software IBM Operating System/2 IBM Business System 12 System/360 OS: IBM Basic Operating System/360 IBM System/360 Operating System IBM Time Sharing System/360 DOS versions: IBM Disk Operating System/360 IBM Tape Operating System/360 IBM 1401 Symbolic Programming System Letter-named series Hardware IBM System i IBM System i5 IBM System p IBM System p5 IBM System x IBM System z IBM System z9 IBM System z10 Software IBM System R IBM SystemT Word-named series Hardware Models and lines: IBM Displaywriter System IBM FlashSystem IBM Flex System IBM Power Systems IBM Personal System/ValuePoint IBM Personal System/V IBM System Storage IBM Q System One IBM XIV Storage System IBM zEnterprise System Technologies: IBM Standard Modular System Software IBM Administrative Terminal System IBM Cross System Product IBM Group Control System IBM Information Management System IBM Internet Security Systems IBM Programming Language/Systems IBM Structured Query Language/Data System IBM Systems Application Architecture IBM Systems Network Architecture IBM System Object Model IBM System Support Program IBM PowerHA SystemMirror File systems: IBM Journaled File System IBM SAN File System See also IBM Series/1 IBM Future Systems project System request button IBM platform (disambiguation) List of IBM products (Operating systems)
16777560
https://en.wikipedia.org/wiki/Chandigarh%20Capital%20Region
Chandigarh Capital Region
Chandigarh Capital Region (CCR) or Chandigarh Metropolitan Region (CMR) is an area, which includes the union territory city of Chandigarh, and its neighboring cities of Mohali, Kharar, Zirakpur, New Chandigarh (in Punjab) and Panchkula (in Haryana). Chandigarh Administration, Greater Mohali Area Development Authority (GMADA) and Haryana Urban Development Authority (HUDA) are different authorities responsible for development of this region. The economy of the region is interdependent as the area is continuously inhabited, though falling under different states. There is lot of movement of people and goods daily to and from suburbs, like most of the people working in Chandigarh live in a suburb like Zirakpur. The local industry is on the outskirts like Derabassi, Lalru and Baddi. History The Indian Ministry of Home Affairs asked the Chandigarh Administration in October 2011 to "coordinate with the Punjab and Haryana governments for working out the modalities" of a Regional Planning Board (RPB) for the Chandigarh Capital Region (CCR). The intention was to harmonize and improve facilities across the region and co-ordinate local administrations. Responsibilities of the RPB would include disaster management, health planning, biomedical waste and traffic management. The 'Master Plan 2031' submitted to the UT administrator in January 2013 included a proposal for a "inter-state regional plan for the Chandigarh Capital Region". However the final plan confined itself to the UT boundary, and did not mention the proposed "Chandigarh Capital Region". Instead, the plan expected responsibilities for the region to be coordinated between Punjab, Haryana and Chandigarh UT, and that the Chandigarh plan "with the GMADA Plan 2056 and the Haryana Development Plan should together arrive at a metropolitan plan". the Chandigarh capital region along the lines of the National Capital Region remains an un-implemented concept. Definition The cities, towns and areas that would be part of Chandigarh Capital region have a total population of 1,758,653 and their city wise population are: Overall Chandigarh ( UT ), Mohali District (Punjab) and Panchkula District ( Haryana ) are part of CCR Derabassi, Lalru, Banur and Kurali in Mohali district. Kalka, Pinjore, Barwala and Raipur Rani in Panchkula district can be considered as another towns in Chandigarh Capital Region In future it can be expanded up to BBN ( Baddi-Brotiwala-Nalagarh) Area in Himachal pradesh and Roopnagar City in Punjab. Industry Chandigarh IT Park is a technology park located within Chandigarh, with presence of companies like IDS Infotech, Airtel, Tech Mahindra and Infosys. Mohali IT City is the infrastructure to facilitate information technology in the city. It spreads over 1700 acres developed by GMADA situated near Chandigarh International Airport. Infosys is given 50 Acres land to develop state of art campus. Quark has 40 Acres campus in Mohali for Quark software Inc. and other IT Companies like Emerson and Infosys. Punwire,Puncom, Mahindra-Swaraj Tractors, Verka, Sun Pharma and Godrej also have Plants in Mohali Dera Bassi - Lalru is another belt with mostly medium industry, this place has many spinning mills including the Nahar group. Bhushan Steel also has a presence here. Panchkula IT Park is the state of art infrastructure to facilitate information technology in the city. It spreads over 74 acres developed by HSIIDC situated in sector 22. Bharat Electronics Limited or B.E.L has a factory in Panchkula. Hindustan Machine Tools has a tractor factory in Pinjore, while Associated Cement Companies has a cement factory in Pinjore. SML-ISUZU has Bus & Truck Plant near Ropar City. This region is the zonal headquarters for a large number of banks, their offices are mainly in the "Bank Square" in sector 17 Chandigarh. This area also houses the regional office (usually covering Punjab, Northern Haryana, Himachal Pradesh and Jammu and Kashmir) for a lot of FMCG companies. Semiconductor laboratory sector 72 mohali. Healthcare Infrastructure The prominent hospitals in the region are Post Graduate Institute of Medical Education and Research PGIMER, Sector 12, Chandigarh Govt Medical College and Hospital (GMCH), Sector 16, Chandigarh Govt Medical College and Hospital (GMCH), Sector 32, Chandigarh Dr. B.R. Ambedkar State Institute of Medical Sciences, Sector 56, Mohali Max Hospital, Sector 56, Mohali Fortis Hospital, Sector 62, Mohali Educational Institutes IIT Ropar IISER Mohali NIPER Mohali ISB Mohali INST Mohali NABI Mohali National Institute of Electronics & Information Technology,Ropar National Institute of Electronics & Information Technology,Chandigarh P.U. Chandigarh Punjab Engineering College, Chandigarh Chandigarh college of Engineering & Technology, Sec-26, Chandigarh PGI Chandigarh IMTECH Chandigarh CSIO Chandigarh TBRL-DRDO Chandigarh GMCH, Chandigarh NIFT Panchkula TATA Memorial Cancer Hospital, Medi City, New Chandigarh Chandigarh University Mohali Plaksha University Mohali Amity University Mohali Army Institute of Law Mohali Sports Punjab Cricket Association IS Bindra Stadium, Mohali Maharaja Yadavindra Singh International Cricket Stadium, New Chandigarh International Hockey Stadium, Mohali GMADA Sports Complex, Sec-78, Mohali Sports Complex, Sector 7, Chandigarh Chandigarh Lawn tennis Association, CLTA Stadium, Sector 10, Chandigarh Cricket Stadium Sector 16, Chandigarh Hockey Stadium and Sports Complex, Sector-42, Chandigarh Golf Course, Sector 6, Chandigarh Tau Devi Lal Cricket Stadium, Sector 3 Panchkula Golf Course, Sector 3 Panchkula Sports Authority of India stadium, Dayalpura, Zirakpur Transport Road transport ISBT, SEC-17, Chandigarh ISBT, SEC-43, Chandigarh ISBT, SEC-57, Mohali Rail transport Chandigarh Railway Station Mohali Railway Station Air transport Chandigarh International Airport is located near Aerocity in Mohali. It has both domestic and international flights. External links Chandigarh Master Plan 2031 References Chandigarh
8312764
https://en.wikipedia.org/wiki/SpySheriff
SpySheriff
SpySheriff is malware that disguises itself as anti-spyware software. It attempts to mislead the user with false security alerts, threatening them into buying the program. Like other rogue antiviruses, after producing a list of false threats, it prompts the user to pay to remove them. The software is particularly difficult to remove, since it nests its components in System Restore folders, and also blocks some system management tools. However, SpySheriff can be removed by an experienced user, antivirus software, or by using a rescue disk. Websites SpySheriff was hosted at both www.spysheriff.com and www.spy-sheriff.com, which operated from 2005 until their shutdown in 2008. Both domains are now parked. Several other similarly-named websites also hosted the program, but have all been shut down. Several typosquatted websites also attempted to automatically install SpySheriff, including a fake version of Google.com called Goggle.com. From 2015 Goggle.com, which had changed ownership following a lawsuit by Google, hosted a survey scam and displayed links to Amazon items. In 2017, the domain hosted a blank page, with only the word "goggle" present in its HTML script. At the beginning of 2018, the site redirected to the scam site tango-deg.com and others, but from October 2018 it existed as a simple HTML markup with a top-level heading reading "Goggle.com Inc.". In late 2019, the website became a WordPress blog, and it is now down. Features of a SpySheriff infection SpySheriff is designed to behave like genuine antispyware software. Its user interface features a progress bar and counts allegedly found threats, but its scan results are deliberately false, with cryptic names such as "Trojan VX …" to mislead and scare the user. Removal attempts may be unsuccessful and SpySheriff may reinstall itself. The desktop background may be replaced with an image resembling a Blue Screen of Death, or a notice reading, "SPYWARE INFECTION! Your system is infected with spyware. Windows recommends that you use a spyware removal tool to prevent loss of data. Using this PC before having it cleaned of spyware threats is highly discouraged." Attempts to remove SpySheriff via Add or Remove Programs in Control Panel either fails or causes the computer to restart unexpectedly. Attempts to connect to the Internet in any Web browser is blocked by SpySheriff. Spy-Sheriff.com becomes the only accessible website, and can be opened through the program's control panel. Attempts to remove SpySheriff via System Restore are blocked as it prevents the calendar and restore points from loading. Users can overcome this by undoing the previous restore operation, after which the system will restore itself, allowing for easier removal of SpySheriff. SpySheriff can detect certain antispyware and antivirus programs running on the machine, and disable them by ending their processes as soon as it detects them. This may prevent its detection and removal by legitimate antivirus programs. SpySheriff can disable Task Manager and Registry Editor, preventing the user from ending its active process or removing its registry entries from Windows. Renaming the 'regedit' and 'taskmgr' executables will solve this problem. See also Rogue security software Trojan horse (computing) Notes References External links http://www.bleepingcomputer.com/forums/topic22402.html https://web.archive.org/web/20120401221555/http://www.microsoft.com/security/portal/Threat/Encyclopedia/Entry.aspx?Name=Program%3AWin32%2FSpySheriff Windows trojans Scareware
53477230
https://en.wikipedia.org/wiki/Troades%20%28Seneca%29
Troades (Seneca)
Troades () is a fabula crepidata (Roman tragedy with Greek subject) of c. 1179 lines of verse written by Lucius Annaeus Seneca. Characters Hecuba, queen of Troy, wife of Priam Chorus of Trojans Talthybius, Greek herald Agamemnon, Greek king Calchas, Greek seer Helena, "Helen of Troy" Pyrrhus, son of Achilles Andromache, wife of Hector senex (old man) Ulysses, Greek Hero Astyanax, son of Andromache and Hector nuntius (messenger) Polyxena, daughter of Hecuba and Priam (silent role) Plot The siege of Troy is done and the city is now smouldering ruins. The victorious Greeks have gathered the rich spoils of Troy upon the shore, among these the Trojan women who await their lot to be assigned to their Greek lords and taken to the cities of their foes. But now the ghost of Achilles has risen from the tomb, and demanded that Polyxena be sacrificed to him before the Greeks shall be allowed to sail away. And Calchas, also, bids that Astyanax be slain, for only then can Greece be safe from any future Trojan war. Act I Hecuba laments with the Chorus of Trojans the destruction of their country and the death of Hector and Priam. Act II Talthybius relates that the Ghost of Achilles has appeared, and reproving the Greeks for their ingratitude, demanded that Polyxena, under the pretext of marriage with whom, he was slain, should be sacrificed at his tomb as an offering to the chthonic gods. Otherwise the Greeks will not have a favourable wind for their return. Agamemnon and Pyrrhus quarrel about the sacrifice. Calchas is summoned and he asserts that not only must Polyxena be slain, but Astynax must also be hurled from the tower. The Chorus denies that Achilles appeared as a spirit, and asserts that the soul dies forever with the body. Act III Andromache having taken alarm at a vision in her dream, hides away her son in his father's tomb. Ulysses in his cleverness discovers where he is, and drags him forth to meet his death. Andromache mingles curses and threats with her supplications entreating Ulysses, but not prevailing upon him. The Trojans, once allotted to the Greeks, are to be conveyed to various parts of Greece—some to Sparta, some to Mycenae, some to Ithaca, and to the country of Helen, Agamemnon and Ulysses. Act IV The plan is discussed as to how the sacrifices to the chthonic gods and manes of Achilles are to be conducted; and in what garments, Polyxena, who is to be sacrificed under the impression of a real marriage, is to be arrayed. Also what part shall be played by Helen, in order that she may cajole Polyxena with the vain hope of marrying Pyrrhus: she at first, keeps up the pretence, but after a time dismisses the deception, having argued with Andromache, she confesses everything and openly recommends the fulfilment of the scheme. The Chorus derives consolation from the misfortune being shared by so many; "as if for the wretched to have companions in sorrow were a solace," and then draws attention to the fact that the solace in question will lose its efficacy, as they will be separated by the allotting that has been going on. Act V The Messenger informs the mothers, Hecuba and Andromache, that Astyanax has been hurled from the tower and Polyxena slain at the tomb of Achilles. Reception Translator R. Scott Smith wrote that Seneca's attempt in the work to weave two episodes together "means that the play is somewhat dissociated – a 'flaw' that critics have sometimes brought to bear against it", but stated that "in the place of unity, however, there is symmetry." References Further reading Elaine Fantham, Seneca’s Troades: A literary introduction with text, translation, and commentary. (Princeton University Press: 1982). Otto Zwierlein (ed.), Seneca Tragoedia (Oxford: Clarendon Press: Oxford Classical Texts: 1986) John G. Fitch Tragedies, Volume I: Hercules. Trojan Women. Phoenician Women. Medea. Phaedra (Cambridge, MA: Harvard University Press: Loeb Classical Library: 2002) Plays by Seneca the Younger Trojan War literature Tragedy plays Agamemnon
188213
https://en.wikipedia.org/wiki/Yahoo%21
Yahoo!
Yahoo (, styled as yahoo!) is an American web services provider. It is headquartered in Sunnyvale, California and operated by the namesake company Yahoo! Inc., which is 90% owned by investment funds managed by Apollo Global Management and 10% by Verizon Communications. It provides a web portal, search engine Yahoo Search, and related services, including My Yahoo!, Yahoo Mail, Yahoo News, Yahoo Finance, Yahoo Sports and its advertising platform, Yahoo! Native. Yahoo was established by Jerry Yang and David Filo in January 1994 and was one of the pioneers of the early Internet era in the 1990s. However, usage declined in the late 2000s as it lost market share to Facebook Groups and Google. History Founding In January 1994, Yang and Filo were electrical engineering graduate students at Stanford University, when they created a website named "Jerry and David's guide to the World Wide Web". The site was a human-edited web directory, organized in a hierarchy, as opposed to a searchable index of pages. In March 1994, "Jerry and David's Guide to the World Wide Web" was renamed "Yahoo!" and became known as the Yahoo Directory. The "yahoo.com" domain was registered on January 18, 1995. The word "yahoo" is a backronym for "Yet Another Hierarchically Organized Oracle" or "Yet Another Hierarchical Officious Oracle". The term "hierarchical" described how the Yahoo database was arranged in layers of subcategories. The term "oracle" was intended to mean "source of truth and wisdom", and the term "officious", rather than being related to the word's normal meaning, described the many office workers who would use the Yahoo database while surfing from work. However, Filo and Yang insist they mainly selected the name because they liked the slang definition of a "yahoo" (used by college students in David Filo's native Louisiana in the late 1980s and early 1990s to refer to an unsophisticated, rural Southerner): "rude, unsophisticated, uncouth." This meaning derives from the Yahoo race of fictional beings from Gulliver's Travels. Yahoo was incorporated on March 2, 1995. In 1995, a search engine function, called Yahoo Search, was introduced. This allowed users to search Yahoo Directory. Yahoo soon became the first popular online directory and search engine on the World Wide Web. Expansion Yahoo grew rapidly throughout the 1990s. Yahoo became a public company via an initial public offering in April 1996 and its stock price rose 600% within two years. Like many search engines and web directories, Yahoo added a web portal, putting it in competition with services including Excite, Lycos, and America Online. By 1998, Yahoo was the most popular starting point for web users, and the human-edited Yahoo Directory the most popular search engine, receiving 95 million page views per day, triple that of rival Excite. It also made many high-profile acquisitions. Yahoo began offering free e-mail from October 1997 after the acquisition of RocketMail, which was then renamed to Yahoo Mail. In 1998, Yahoo replaced AltaVista as the crawler-based search engine underlying the Directory with Inktomi. Yahoo's two biggest acquisitions were made in 1999: Geocities for $3.6 billion and Broadcast.com for $5.7 billion. Its stock price skyrocketed during the dot-com bubble, closing at an all-time high of $118.75/share on January 3, 2000. However, after the dot-com bubble burst, it reached a post-bubble low of $8.11 on September 26, 2001. Yahoo began using Google for search in June 2000. Over the next four years, it developed its own search technologies, which it began using in 2004 partly using technology from its $280 million acquisition of Inktomi in 2002. In response to Google's Gmail, Yahoo began to offer unlimited email storage in 2007. In 2008, the company laid off hundreds of people as it struggled from competition. In February 2008, Microsoft made an unsolicited bid to acquire Yahoo for $44.6 billion. Yahoo rejected the bid, claiming that it "substantially undervalues" the company and was not in the interest of its shareholders. Although Microsoft increased its bid to $47 billion, Yahoo insisted on another 10%+ increase to the offer and Microsoft cancelled the offer in May 2008. Carol Bartz, who had no previous experience in Internet advertising, replaced Yang as CEO in January 2009. In September 2011, after failing to meet targets, she was fired by chairman Roy J. Bostock; CFO Tim Morse was named as Interim CEO of the company. In April 2012, after the appointment of Scott Thompson as CEO, several key executives resigned, including chief product officer Blake Irving. On April 4, 2012, Yahoo announced 2,000 layoffs, or about 14% of its 14,100 workers by the end of year, expected to save around $375 million annually. In an email sent to employees in April 2012, Thompson reiterated his view that customers should come first at Yahoo. He also completely reorganized the company. On May 13, 2012, Thompson was fired and was replaced on an interim basis by Ross Levinsohn, recently appointed head of Yahoo's new Media group. Several associates of Third Point Management, including Daniel S. Loeb were nominated to the board of directors. Thompson's total compensation for his 130-day tenure with Yahoo was at least $7.3 million. On July 15, 2012, Marissa Mayer was appointed president and CEO of Yahoo, effective July 17, 2012. In June 2013, Yahoo acquired blogging site Tumblr for $1.1 billion in cash, with Tumblr's CEO and founder David Karp continuing to run the site. In July 2013, Yahoo announced plans to open an office in San Francisco. On August 2, 2013, Yahoo acquired Rockmelt; its staff was retained, but all of its existing products were terminated. Data collated by comScore during July 2013 revealed that, during the month, more people in the U.S. visited Yahoo websites than Google; the first time that Yahoo outperformed Google since 2011. The data did not count mobile usage, nor Tumblr. Mayer also hired Katie Couric to be the anchor of a new online news operation and started an online food magazine. However, by January 2014, doubts about Mayer's progress emerged when Mayer fired her own first major hire, Henrique de Castro. On December 12, 2014, Yahoo acquired video advertising provider BrightRoll for $583 million. On November 21, 2014, Yahoo acquired Cooliris. Decline, security breaches, and sale By December 2015, Mayer was criticized as performance declined. Mayer was ranked as the least likable CEO in tech. On February 2, 2016, Mayer announced layoffs amounting to 15% of the Yahoo! workforce. On July 25, 2016, Verizon Communications announced the acquisition of Yahoo's core Internet business for $4.83 billion. The deal excluded Yahoo's 15% stake in Alibaba Group and 35.5% stake in Yahoo Japan. On February 21, 2017, as a result of the Yahoo data breaches, Verizon lowered its purchase price for Yahoo by $350 million and reached an agreement to share liabilities regarding the data breaches. On June 13, 2017, Verizon completed the acquisition of Yahoo and Marissa Mayer resigned. Yahoo, AOL, and HuffPost were to continue operating under their own names, under the umbrella of a new company, Oath Inc., later called Verizon Media. The parts of the original Yahoo! Inc. which were not purchased by Verizon Communications were renamed Altaba, which later liquidated, making a final distribution in October 2020. In September 2021, investment funds managed by Apollo Global Management acquired 90% of Yahoo. In November 2021, Yahoo announced that it was ceasing its operations in mainland China due to an increasingly challenging business and legal environment. Chief Executive Officers Eleven chief executives and interim leaders have led the Yahoo companies since 1995. They are: Jim Lanzone, CEO of Yahoo Inc. (2021–present) Guru Gowrappan, CEO of Oath Inc., Verizon Media, and Yahoo (2018–2021) Tim Armstrong, CEO of Oath Inc. (2017–2018) Marissa Mayer (2012–2017) Ross Levinsohn Interim (2012) Scott Thompson (2012) Tim Morse Interim (2011–2012) Carol Bartz (2009–2011) Jerry Yang (2007–2009) Terry Semel (2001–2007) Timothy Koogle (1995–2001) Products and services For a list of all current and defunct services offered by Yahoo, see List of Yahoo-owned sites and services. Data breaches On September 22, 2016, Yahoo disclosed a data breach that occurred in late 2014, in which information associated with at least 500 million user accounts, one of the largest breaches reported to date. The United States indicted four men, including two employees of Russia's Federal Security Service (FSB), for their involvement in the hack. On December 14, 2016, the company revealed that another separate data breach had occurred in 2014, with hackers obtaining sensitive account information, including security questions, to at least one billion accounts. The company stated that hackers had utilized stolen internal software to forge HTTP cookies. On October 3, 2017, the company stated that all 3 billion of its user accounts were affected by the August 2013 theft. Criticism DMCA notice to whistleblower On November 30, 2009, Yahoo was criticized by the Electronic Frontier Foundation for sending a DMCA notice to whistleblower website "Cryptome" for publicly posting details, prices, and procedures on obtaining private information pertaining to Yahoo's subscribers. Censorship of private emails affiliated with Occupy Wall Street protests After some concerns over censorship of private emails regarding a website affiliated with Occupy Wall Street protests were raised, Yahoo responded with an apology and explained it as an accident. Sponsorship On September 11, 2001, Yahoo! announced its partnership with FIFA for the 2002 FIFA World Cup and 2006 FIFA World Cup tournaments. It was one of FIFA's 15 partners at the tournaments. The deal included co-branding the organization's websites. Yahoo! sponsored the 2012 Sundance Film Festival. Logos and themes The first Yahoo logo was established upon foundation of the company in 1994. It consists of the "Yahoo" wordmark which is colored black and is using the Times New Roman font, but it was later changed. In March 1995, when the company changed its name to Yahoo, it introduced another logo which is briefly changed to a more elaborate text that includes an exclamation point at the end. In August 1995, Yahoo changed its logo to a stylized yellow jumping "Y" figurine on a blue circle while the "Yahoo!" wordmark is written below. On January 1, 1996, Yahoo introduced a simplified new logo that included the text "Yahoo" and an exclamation mark, both in red with a slight shadow behind the text. By May 2009, Yahoo tweaked the logo by recoloring it from red to purple and removing the logo's outline and shadow. At the time, the purple logo was accompanied by a new slogan, "It's Y!ou." A shortened variant of the logo, consisting of only the letter "Y" and an exclamation point, was also used. On August 7, 2013, at around midnight EDT, Yahoo announced that the final version of the new logo would be revealed on September 5, 2013, at 4:00 a.m. UTC. In the period leading up to the unveiling of the new logo, the "30 Days of Change" campaign was introduced, whereby a variation of the logo was published every day for the 30 days following the announcement. The new logo was eventually launched with an accompanying video that showed its digital construction, and Mayer published a personalized description of the design process on her Tumblr page. Mayer explains: So, one weekend this summer, I rolled up my sleeves and dove into the trenches with our logo design team ... We spent the majority of Saturday and Sunday designing the logo from start to finish, and we had a ton of fun weighing every minute detail. We knew we wanted a logo that reflected Yahoo – whimsical, yet sophisticated. Modern and fresh, with a nod to our history. Having a human touch, personal. Proud. On September 19, 2013, Yahoo launched a new version of the "My Yahoo" personalized homepage. The redesign allows users to tailor a homepage with widgets that access features such as email accounts, calendars, Flickr and other Yahoo content, and Internet content. Users can also select "theme packs" that represent artists such as Polly Apfelbaum and Alec Monopoly, and bands such as Empire of the Sun. Mayer then explained at a conference in late September 2013 that the logo change was the result of feedback from both external parties and employees. In September 2019 Yahoo changed its logo again for a "refreshed brand identity [that] is simpler and more flexible, and looks back to the original, quirky 1996 logo." The logo is a white text set against a purple background, with both the “y” and “!” of the logo reportedly set at an angle of 22.5 degrees. The logo was designed by Pentagram. See also List of Yahoo!-owned sites and services List of search engines Yahoo! litigation References External links Yahoo! Products and Services ALTABA and Yahoo! EDGAR Filing History Companies based in Sunnyvale, California Companies in the PRISM network Companies formerly listed on the Nasdaq Internet properties established in 1994 Multinational companies headquartered in the United States Software companies based in California Technology companies based in the San Francisco Bay Area Technology companies established in 1994 Telecommunications companies established in 1994 1994 establishments in California 2017 mergers and acquisitions Software companies of the United States
8293945
https://en.wikipedia.org/wiki/Vuze
Vuze
Vuze (previously Azureus) is a BitTorrent client used to transfer files via the BitTorrent protocol. Vuze is written in Java, and uses the Azureus Engine. In addition to downloading data linked to .torrent files, Azureus allows users to view, publish and share original DVD and HD quality video content. Content is presented through channels and categories containing TV shows, music videos, movies, video games, series and others. Additionally, if users prefer to publish their original content, they may earn money from it. Azureus was first released in June 2003 at SourceForge.net, mostly to experiment with the Standard Widget Toolkit from Eclipse. It later became one of the most popular BitTorrent clients. The Azureus software was released under the GNU General Public License, and remains as a free software application. It was among the most popular BitTorrent clients. However, the Vuze software added in more recent versions is proprietary and users are required to accept these more restrictive license terms. Features Bridging between I2P and the clearnet Vuze is the only client that makes clearnet torrents available on I2P and vice versa. It has a plugin that connects to the I2P network. If the user adds a torrent from I2P, it will be seeded on both I2P and the clearnet, and if a user adds a torrent from the clearnet, it will be seeded on both the clearnet and I2P. For this reason, torrents previously published only on I2P are made available to the entire Internet, and users of I2P can download any torrent on the Internet while maintaining the anonymity of I2P. History Azureus Azureus was first released in June 2003 at SourceForge.net. The blue poison dart frog (Dendrobates azureus) was chosen as the logo and name of the brand by co-creator Tyler Pitchford. This choice was due to Latin names of poison dart frogs being used as codenames for his development projects. Vuze In 2006 Vuze was released as an attempt to transform the client into a "social" client by a group of the original developers forming Azureus Inc., shortly to be renamed Vuze, Inc. A Vuze-free version of Azureus was released along with Vuze during the beta period. The releases used version numbers 3.0, while the Vuze-free versions continued with the 2.5 release numbers. In addition, some developers voiced opposition to the idea of completely transforming the client. Starting with an unknown version, Vuze was coupled with Azureus. Soon after, "NoVuze" modified versions were released on The Pirate Bay, and as of September 15, 2008, are available for versions up to 3.1.1.0. On June 16, 2008, the developers of Azureus/Vuze decided to stop releasing versions named Azureus, and complete the name change with the release of version 3.1. The client engine however, remains unchanged as Azureus. License change Up to version 2.5.0.4, Azureus was distributed under the GNU General Public License (GPL); beginning with the version 3 distribution, the license presented upon installation changed. While it still states that the "Azureus Application" is available under the GPL, completing installation requires the user to agree to the terms of the "Vuze Platform," which include restrictions on use, reverse-engineering, and sublicensing. As with many similar licenses, the Azureus licence includes a prohibition on use of the software by people "under the age of 18." Allegedly, the TOS only applies to the website, vuze.com, and not the software, however the actual TOS include the application as part of the platforms. Acquisition by Spigot Inc Vuze was acquired by Spigot Inc in December 2010. Development hiatus The pace of development slowed in early 2017. the latest entry in the development blog is dated April 2017 and the latest code change is dated February 2018. Criticism Vuze is categorized as adware by Softpedia, due to its inclusion of a Vuze Toolbar for web browsers. Vuze changes or offers to change home page and search and to install a promotional component not necessary for the program to function. However, all adware can be declined by using a custom installation. In February 2010, What.CD and Waffles.fm, two large music sharing sites at the time, decided to ban the use of Vuze. However, Vuze cites its Softpedia Editor's pick award, having received an editor score of 4 out of 5 from two reviews: One on 23 November 2005 and another on 7 February 2012. Vuze includes built-in support for Tor, an anonymity network. The onion routers are run by volunteers using their own bandwidth at their own cost. Due to the high bandwidth usage caused by the BitTorrent protocol, it is considered impolite and inappropriate by Tor community members to use the Tor network for BitTorrent transfers. By default, the Tor exit policy blocks the standard BitTorrent ports. It also includes I2P support via an official plugin. In contrast to Tor, I2P is built for P2P traffic and encourages its use. See also Comparison of BitTorrent clients Spigot Inc References External links Android (operating system) software 2003 software Java platform software MacOS file sharing software BitTorrent clients for Linux Windows file sharing software
31331954
https://en.wikipedia.org/wiki/List%20of%20Lepidoptera%20of%20Cape%20Verde
List of Lepidoptera of Cape Verde
Lepidoptera of Cape Verde represent about 280 known species. A total of 252 moth species have been recorded. Twenty-eight species of butterflies are known from Cape Verde, none of which are endemic. The moths (mostly nocturnal) and butterflies (mostly diurnal) together make up the taxonomic order Lepidoptera. Butterflies Hesperiidae Borbo borbonica (Boisduval, 1833) Coeliades forestan (Stoll, 1782) Lycaenidae Azanus jesous (Guérin, 1847) Azanus mirza (Plötz, 1880) Azanus moriqua (Wallengren, 1857) Chilades evorae Libert, Baliteau & Baliteau, 2011 Deudorix dinomenes Grose-Smith, 1887 Euchrysops osiris (Hopffer, 1855) Lampides boeticus (Linnaeus, 1767) Leptotes pirithous (Linnaeus, 1767) Zizeeria knysna (Trimen, 1862) Nymphalidae Byblia ilithyia (Drury, 1773) Danaus chrysippus (Cramer, 1777) Nymphalis polychloros (Linnaeus, 1758) Hypolimnas misippus(Linnaeus, 1764) Vanessa cardui (Linnaeus, 1758) Papilionidae Papilio demodocus Esper, 1798 Pieridae Belenois creona (Cramer, 1776) Catopsilia florella (Fabricius, 1775) Colias croceus (Geoffroy in Fourcroy, 1785) Colotis amata (Fabricius, 1775) Colotis euippe (Linnaeus, 1758) Eurema floricola (Boisduval, 1833) Eurema hecabe (Linnaeus, 1758) Eurema senegalensis (Boisduval, 1836) Pontia daplidice (Linnaeus, 1758) Pontia glauconome Klug, 1829 Satyrinae Melanitis leda (Linnaeus, 1758) Moths Bedelliidae Bedellia somnulentella (Zeller, 1847) Coleophoridae Coleophora creola Baldizzone & van der Wolf, 2015 Cosmopterigidae Anatrachyntis simplex (Walsingham, 1891) Cosmopterix attenuatella (Walker, 1864) Crambidae Achyra coelatalis (Walker, 1859) Achyra nudalis (Hübner, 1796) Cataonia mauritanica Amsel, 1953 Condylorrhiza zyphalis (Viette, 1958) Cornifrons ulceratalis Lederer, 1858 Cynaeda dentalis (Denis & Schiffermüller, 1775) Diaphania indica (Saunders, 1851) Duponchelia fovealis Zeller, 1847 Euclasta insularis Viette, 1958 Eudonia fogoalis Derra, 2008 Eudonia lindbergalis Viette, 1958 Hellula undalis (Fabricius, 1781) Herpetogramma licarsisalis (Walker, 1859) Hydriris ornatalis (Duponchel, 1832) Cnaphalocrocis poeyalis (Boisduval, 1833) Maruca vitrata (Fabricius, 1787) Nomophila noctuella (Denis & Schiffermüller, 1775 Notarcha quaternalis (Zeller, 1852) Omiodes indicata (Fabricius, 1775) Orphanostigma abruptalis (Walker, 1859) Palpita vitrealis (Rossi, 1794) Pediasia strenua Bassi, 1992 Spoladea recurvalis (Fabricius, 1775) Udea ferrugalis (Hübner, 1796) Depressariidae Ethmia quadrinotella (Mann, 1861) Erebidae Acantholipes aurea Berio, 1966 Acantholipes trimeni Felder & Rogenhofer, 1875 Achaea finita (Guenée, 1852) Achaea infinita (Guenée, 1852) Achaea violaceofascia (Saalmüller, 1891) Anomis auragoides (Guenée, 1852) Anomis flava (Fabricius, 1775) Anticarsia rubricans (Boisduval, 1833) Asota speciosa (Drury, 1773) Audea melaleuca Walker, 1865 Crypsotidia maculifera (Staudinger, 1898) Crypsotidia mesosema Hampson, 1913 Crypsotidia remanei Wiltshire, 1977 Dysgonia algira (Linnaeus, 1767) Dysgonia angularis (Boisduval, 1833) Dysgonia torrida (Guenée, 1852) Eilema aistleitneri Cerny, 2013 Ericeia congregata (Walker, 1858) Ericeia inangulata (Guenée, 1852) Eublemma apicimacula (Mabille, 1880) Eublemma baccatrix Hacker, 2019 Eublemma cochylioides (Guenée, 1852) Eublemma ecthaemata Hampson, 1896 Eublemma exigua (Walker, 1858) Eublemma gayneri (Rothschild, 1901) Eublemma mesophaea Hampson, 1910 Eublemma parva (Hübner, 1808) Eublemma scitula (Rambur, 1833) Eublemma spirogramma Rebel, 1912 Gesoniodes nigripalpa (Wiltshire, 1977) Gnamptonyx innexa (Walker, 1858) Grammodes congenita Walker, 1858 Grammodes exclusiva Pagenstecher, 1907 Grammodes stolida (Fabricius, 1775) Hipoepa fractalis (Guenée, 1854) Hypena abyssinialis Guenée, 1854 Hypena conscitalis Walker, 1866 Hypena laceratalis Walker, 1859 Hypena lividalis (Hübner, 1790) Hypena semilutea (Snellen, 1872) Hypena strigatus (Fabricius, 1798) Hypena varialis Walker, 1866 Janseodes melanospila (Guenée, 1852) Lygephila pastinum (Treitschke, 1826) Maxera marchalii (Boisduval, 1833) Maxera nigriceps (Walker, 1858) Mesogenea varians Hampson, 1902 Mocis conveniens (Walker, 1858) Mocis mayeri (Boisduval, 1833) Mocis proverai Zilli, 2000 Nodaria externalis Guenée, 1854 Ophiusa tirhaca (Cramer, 1777) Oraesia intrusa (Krüger, 1939) Pandesma robusta (Walker, 1858) Pericyma mendax (Walker, 1858) Phytometra melanosticta Hacker, 2016 Polydesma umbricola Boisduval, 1833 Serrodes partita (Fabricius, 1775) Simplicia extinctalis (Zeller, 1852) Sphingomorpha chlorea (Cramer, 1777) Tathorhynchus exsiccata (Lederer, 1855) Tathorhynchus leucobasis Bethune-Baker, 1911 Tathorhynchus troberti (Guenée, 1852) Trigonodes hyppasia (Cramer, 1779) Ulotrichopus primulina (Hampson, 1902) Utetheisa pulchella (Linnaeus, 1758) Gelechiidae Anarsia balioneura Meyrick, 1921 Aristotelia benedenii (Weyenbergh, 1873) Brachmia convolvuli (Walsingham, 1907) Helcystogramma lamprostoma (Zeller, 1847) Ornativalva heluanensis (Debski, 1913) Pectinophora gossypiella (Saunders, 1844) Phthorimaea operculella (Zeller, 1873) Scrobipalpa aptatella (Walker, 1864) Zizyphia cleodorella Chrétien, 1908 Geometridae Chiasmia sudanata (Warren & Rothschild, 1905) Comibaena leucospilata (Walker, 1863) Comostolopsis stillata (Felder & Rogenhofer, 1875) Eucrostes beatificata (Walker, 1863) Eucrostes disparata Walker, 1861 Gymnoscelis daniloi Hausmann, 2009 Gymnoscelis lindbergi Herbulot, 1957 Isturgia catalaunaria (Guenée, 1858) Isturgia deerraria (Walker, 1861) Mesocolpia nanula (Mabille, 1900) Microloxia aistleitneri Hausmann, 2009 Microloxia herbaria (Hübner, 1818) Microloxia ruficornis Warren, 1897 Orthonama obstipata (Fabricius, 1794) Pasiphila derasata (Bastelberger, 1905) Phaiogramma faustinata (Millière, 1868) Pingasa hypoleucaria (Guenée, 1862) Pingasa lahayei (Oberthür, 1887) Pingasa rhadamaria (Guenée, 1858) Prasinocyma germinaria (Guenée, 1857) Scopula minorata (Boisduval, 1833) Scopula paneliusi Herbulot, 1957 Scopula retracta (Hausmann, 2006) Thalassodes quadraria Guenée, 1858 Traminda vividaria (Walker, 1861) Gracillariidae Caloptilia soyella (van Deventer, 1904) Phodoryctis caerulea (Meyrick, 1912) Noctuidae Acontia basifera Walker, 1857 Acontia conifrons Aurivillius, 1879 Acontia feae (Berio, 1937) Acontia gratiosa Wallengren, 1856 Acontia imitatrix Wallengren, 1856 Acontia insocia (Walker, 1858) Acontia opalinoides Guenée, 1852 Adisura callima Bethune-Baker, 1911 Agrotis aistleitneri Behounek & Speidel, 2009 Agrotis ipsilon (Hufnagel, 1766) Agrotis segetum (Denis & Schiffermüller, 1775) Agrotis spinifera (Hübner, 1808) Agrotis subspinifera (Hampson, 1903) Agrotis trux (Hübner, 1824) Amyna axis Guenée, 1852 Anarta trifolii (Hufnagel, 1766) Androlymnia clavata Hampson, 1910 Asplenia melanodonta (Hampson, 1896) Athetis ochreosignata Aurivillius, 1910 Athetis pigra (Guenée, 1852) Callopistria latreillei (Duponchel, 1827) Callopistria maillardi (Guenée, 1862) Caradrina clavipalpis (Scopoli, 1763) Caradrina flava Oberthür, 1876 Chasmina tibialis (Fabricius, 1775) Chasmina vestae (Guenée, 1852) Chrysodeixis acuta (Walker, 1858) Chrysodeixis chalcites (Esper, 1798) Condica capensis (Guenée, 1852) Condica conducta (Walker, 1857) Condica pauperata (Walker, 1858) Cornutiplusia circumflexa (Linnaeus, 1767) Ctenoplusia limbirena (Guenée, 1852) Euxoa admirabilis Hacker & Schreier, 2010 Euxoa canariensis Rebel, 1902 Haplocestra similis Aurivillius, 1910 Helicoverpa armigera (Hübner, 1808) Helicoverpa assulta (Guenée, 1852) Heliocheilus confertissima (Walker, 1865) Heliothis nubigera Herrich-Schäffer, 1851 Heliothis peltigera (Denis & Schiffermüller, 1775) Hiccoda pluristriata (Berio, 1937) Iambiodes incerta (Rothschild, 1913) Leucania loreyi (Duponchel, 1827) Maliattha sahelica Hacker, 2016 Maliattha signifera (Walker, 1857) Mythimna languida (Walker, 1858) Mythimna natalensis (Butler, 1875) Mythimna poliastis (Hampson, 1902) Mythimna umbrigera (Saalmüller, 1891) Mythimna vilis (Gaede, 1916) Oederemia simplivalva Hacker & Schreier, 2010 Ozarba adducta Berio, 1940 Ozarba exoplaga Berio, 1940 Ozarba phlebitis Hampson, 1910 Ozarba rubrivena Hampson, 1910 Pardoxia graellsii (Feisthamel, 1837) Peridroma saucia (Hübner, 1808) Pseudozarba bipartita (Herrich-Schäffer, 1850) Pseudozarba expatriata (Hampson, 1914) Pseudozarba opella (Swinhoe, 1885) Saalmuellerana media (Walker, 1857) Sesamia calamistis Hampson, 1910 Sesamia nonagrioides (Lefèbvre, 1827) Spodoptera exempta (Walker, 1857) Spodoptera exigua (Hübner, 1808) Spodoptera littoralis (Boisduval, 1833) Thysanoplusia orichalcea (Fabricius, 1775) Trichoplusia ni (Hübner, 1803) Vittaplusia vittata (Wallengren, 1856) Xanthodes albago (Fabricius, 1794) Nolidae Earias biplaga Walker, 1866 Earias cupreoviridis (Walker, 1862) Earias insulana (Boisduval, 1833) Pardasena virgulana (Mabille, 1880) Plutellidae Plutella xylostella (Linnaeus, 1758) Praydidae Prays citri (Millière, 1873) Pterophoridae Agdistis bifurcatus Agenjo, 1952 Agdistis notabilis Gielis & Karsholt, 2009 Agdistis tamaricis (Zeller, 1847) Exelastis atomosa (Walsingham, 1885) Exelastis pumilio (Zeller, 1873) Hellinsia aistleitneri Arenberger, 2006 Lantanophaga pusillidactylus (Walker, 1864) Megalorhipida leucodactylus (Fabricius, 1794) Sphenarches anisodactylus (Walker, 1864) Stenoptilodes taprobanes (Felder & Rogenhofer, 1875) Pyralidae Cadra figulilella (Gregson, 1871) Corcyra cephalonica (Stainton, 1866) Cryptoblabes gnidiella (Millière, 1867) Ephestia elutella (Hübner, 1796) Etiella zinckenella (Treitschke, 1832) Galleria mellonella (Linnaeus, 1758) Phycita pachylepidella Hampson, 1896 Plodia interpunctella (Hübner, 1813) Susia uberalis (Swinhoe, 1884) Thylacoptila paurosema Meyrick, 1885 Scythrididae Eretmocera basistrigata Walsingham, 1889 Eretmocera laetissima Zeller, 1852 Scythris fissurella Bengtsson, 1997 Sphingidae Acherontia atropos (Linnaeus, 1758) Agrius cingulata (Fabricius, 1775) Agrius convolvuli (Linnaeus, 1758) Basiothia medea (Fabricius, 1781) Daphnis nerii (Linnaeus, 1758) Hippotion celerio (Linnaeus, 1758) Hyles euphorbiae (Linnaeus, 1758) Hyles livornica (Esper, 1780) Hyles tithymali (Boisduval, 1834) Nephele accentifera (Palisot de Beauvois, 1821) Tineidae Opogona sacchari (Bojer, 1856) Tortricidae Ancylis lutescens Meyrick, 1912 Coniostola stereoma (Meyrick, 1912) Cydia choleropa (Meyrick, 1913) Eccopsis praecedens Walsingham, 1897 Eccopsis wahlbergiana Zeller, 1852 Fulcrifera boavistae Razowski, 2015 Strepsicrates rhothia (Meyrick, 1910) Thaumatotibia leucotreta (Meyrick, 1913) References B Lepidoptera Cape Verde Cape Verde Cape Verde
66019828
https://en.wikipedia.org/wiki/OneFuzz
OneFuzz
OneFuzz is a cross-platform free and open source fuzz testing framework by Microsoft. The software enables continuous developer-driven fuzz testing to identify weaknesses in computer software prior to release. Overview OneFuzz is a self-hosted fuzzing-as-a-service platform that automates the detection of software bugs that could be security issues. It supports Windows and Linux. Notable features include composable fuzzing workflows, built-in ensemble fuzzing, programmatic triage and result de-duplication, crash reporting notification callbacks, and on-demand live-debugging of found crashes. The command-line interface client is written in Python 3, and targets Python 3.7 and up. Microsoft uses the OneFuzz testing framework to probe Edge, Windows and other products at the company. It replaced the previous Microsoft Security Risk Detection software testing mechanism. The source code was released on September 18, 2020. It is licensed under MIT License and hosted on GitHub. See also Test automation Random testing American fuzzy lop (fuzzer) DynamoRIO Pin (computer program) References External links Microsoft announces new Project OneFuzz framework, an open source developer tool to find and fix bugs at scale Free and open-source software Free software programmed in Rust Free software testing tools Security testing tools Microsoft free software Microsoft Research Software using the MIT license 2020 software
34776878
https://en.wikipedia.org/wiki/Stencyl
Stencyl
Stencyl is a video game development tool that allows users to create 2D video games for computers, mobile devices, and the web. The software is available for free, with select publishing options available for purchase. The software was originally called "StencylWorks" while in development and for the initial release but was later shortened to just "Stencyl". Features Games created in Stencyl can be exported to the web via Adobe Flash Player, and to personal computers as executable games, as well as onto various mobile devices as iOS and Android applications. Physics and collisions are managed by Box2D, which can be selectively or completely disabled to decrease any potential performance impact for games that don't require full physics simulation. Starting in version 3.0, projects in Stencyl use the Haxe programming language and OpenFL game framework to allow a flexible, write once, run anywhere style of game creation. IDE Stencyl is an authoring tool and an IDE. The application includes several modules used to accomplish the necessary tasks to create games with the software. The Behavior Editor is used to create and edit code and game logic in modular pieces known as behaviours and events. The Tileset Editor is used to import and edit tilesets, including their collision shapes, appearance, and animations. The Actor Editor is used to create and edit game entities (Actors) and their settings, including behaviours, physics, and animations. The Scene Designer is used to create and edit levels and game states (Scenes) by using actors, tilesets, and behaviours. Additional tools permit the user to import images for use as foregrounds and backgrounds in scenes, import and edit fonts, import sounds and music files (MP3 and OGG are supported, depending on the export target), and alter game settings such as player controls and game resolution. A library of common behaviours is included with Stencyl to reduce the need to recreate common game behaviours, and several game "kits" provide functional starting points for common 2D game genres. VPL When creating a new behaviour, the option is presented to create it either in Code Mode or Design Mode. Using Code Mode for a behaviour permits the user to program logic in traditional textual form and optionally open the code in an external editor. Alternatively, Design Mode is a GUI that allows users to create modular game logic for actors and scenes using a visual programming language. The concept of Design Mode as a form of end-user development originated with MIT's Scratch computer language learning environment, and was used with permission for Stencyl. As it is a visual programming language, Design Mode users are not required to learn or type out a particular programming language, nor must they concern themselves with syntax. Rather, available actions are dragged and dropped from a palette of "code blocks". These blocks will snap in place together and nest within each other, permitting the creation of advanced logic from basic components. To avoid syntax errors during compilation, not all blocks will snap together. For example, a space that requires a Boolean value will not accept a block that represents a numeric value. The shapes of the block types are different to help represent this to the user as a behaviour-shaping constraint. The number block could be used in conjunction with a comparison block - such as "(Number) equals (Number)" - to evaluate as a True/False statement for the needed Boolean. Cloud Storage StencylForge is an integrated online storage and sharing service for Stencyl games and game assets. It can be used to back up projects and access projects from other computers when logged into Stencyl. User-uploaded content is private by default, but it can also be made public to share with other users. In addition to user-uploaded content, officially sanctioned content such as example games, game kits, behaviours, and art or sound assets are also available and prominently featured on StencylForge. Other Integrated with Stencyl is Pencyl, an image editor derived from JHLabs Image Editor. Its use in Stencyl is usually secondary, as images are typically imported directly from the computer after being created separately. Stencyl can alternatively be set up to use external image editors, such as Photoshop and GIMP, to modify images already loaded into a project. Stencyl also supports user-created extensions to add functions to the software. For example, an extension could add new blocks for Design Mode use, such as a third-party API. Another example is an extensive dialogue scripting tool to easily add dialogue and tweak its settings for a game. See also AgentCubes AgentSheets Scratch (programming language) GameSalad GameMaker: Studio Construct Corona (software development kit) RPG Maker References External links Official website Video game IDE Video game engines Game engines for Linux Video game development software Visual programming languages Educational programming languages Browser-based game websites Video game development software for Linux
56736105
https://en.wikipedia.org/wiki/Christopher%20Macosko
Christopher Macosko
Christopher W. Macosko (1944) is an American chemical engineer and professor emeritus in the Department of Chemical Engineering and Materials Science at the University of Minnesota. He is internationally known for his work in polymer science and engineering, especially in the areas of rheology and polymer processing. Macosko is an author of more than 500 academic papers, dozens of patents, and two books including the text: "Rheology: Principles, Measurements and Applications" (Wiley/VCH, 1994). He served as Director of the Industrial Partnership for Research in Interfacial and Materials Engineering (IPRIME), a university-industry consortium at the University of Minnesota, from 1999 to 2018. Macosko and his wife Kathleen have been married since 1967 and are long-time residents of Minneapolis. They have four children and 12 grandchildren. Early life and education Macosko received a B.S. degree in Chemical Engineering in 1966 from Carnegie-Mellon University. He then spent one year at Imperial College, London and obtained an M.S. degree in Chemical Engineering in 1967. Macosko earned his Ph.D. degree in Chemical Engineering in 1970 from Princeton University, where he worked under the supervision of Bryce Maxwell. His thesis work included the design of a new less-compliant rheometer. Macosko and a fellow graduate student, Joe Starita, co-founded the company Rheometrics, whose instruments have significantly advanced the field of rheology. The company is now part of TA Instruments, a world-leading manufacturer of rheological devices. Research at Minnesota In 1970, Macosko joined the University of Minnesota and become one of the first assistant professors in the then newly combined Department of Chemical Engineering and Materials Science. His research on polymers served as a bridge between the two disciplines, and his leadership helped attract other faculty in this area and contributed to the international reputation in polymers at the University of Minnesota. Macosko's research was centered around polymer processing, with special emphasis on the structure-function relation between polymer morphology and the associated rheological, mechanical, electronic, and other physical properties. He made significant contributions toward understanding the rheology of polymer materials, such as the characterization of network formation during gelation, the elasticity of foams, interfacial area generation in reactive blends, and the viscoelastic character of polymer composites. Macosko's research has addressed a broad range of polymer processing operations including reaction injection molding, extrusion, melt blowing, and coating. His current research interests include network polymerization, nanoparticle reinforced polymer composites, interfacial reaction during polymer blending, and polymerization with phase separation. Macosko has taught short courses for many years on rheological measurements that have attracted industrial interest. Awards, honors, and professional service , Macosko has graduated over 100 Ph.D. students, and has supervised more than 50 postdoctoral researchers. Many of his students and postdocs have become faculty members at major research universities across the world (e.g., Julio Ottino at Northwestern University). Macosko was elected a member of the National Academy of Engineering in 2001 for the invention, development, and dissemination of new methods of reactive polymer processing and rheological property measurement. He was awarded the Bingham Medal by the Society of Rheology in 2004. He has also received Charles M. A. Stine Award (now called the Braskem Award) in Materials Engineering and Sciences from American Institute of Chemical Engineers in 1988 at Northwestern University. In addition, Macosko has received numerous awards including the Pall Award for Applied Polymer Research in 1997, the Banbury Award from American Chemical Society in 2006, the International Award from the Society of Polymer Science Japan (SPSJ) in 2008, and the Lifetime Achievement Award from the International Conference on Advanced Molding Technology and Materials Processing in 2012. Works Chris Macosko has authored numerous journal articles describing significant advances in polymer rheology, melt processing, and nanocomposites which includes but is not limited to: C.W. Macosko, D.R. Miller "A new derivation of postgel properties of network polymers", Rubber Chemistry and Technology, 49(5), 1219-1231, (1976). C.W. Macosko, D.R. Miller "A new derivation of average molecular weights of nonlinear polymers", Macromolecules, 9(2), 199-206, (1976). A.C. Papanastasiou, L.E. Scriven, C.W. Macosko "An Integral Constitutive Equation for Mixed Flows: Viscoelastic Characterization", Journal of Rheology, 27(4), 387-410, (1983). D.E. Bornside, C.W. Macosko, L.E. Scriven "Spin Coating: One-dimensional Model", Journal of Applied Physics, 66(11), 5185-5193, (1989). U. Sundararaj, C.W. Macosko "Drop breakup andcoalescence in polymer blends: the effects of concentration and compatibilization", Macromolecules, 38(8), 2647-2657, (1995). X. Cao, L.J. Lee, T. Widya, C.W. Macosko "Polyurethane/clay nanocomposite foams: processing, structure and properties", Polymer, 46(3), 775-783, (2005). C.J. Ellison, A. Phatak, D.W. Giles, C.W. Macosko, F.S.Bates "Melt blown nanofibers: fiber diameter distributions and onset of fiber breakup", Polymer, 48(11), 3306-3316, (2007). H. Kim, A.A. Abdala, C.W. Macosko "Graphene/Polymer Nanocomposites", Macromolecules, 43(16), 6515-6530, (2010). Chris was also the author of two highly influential books: C.W. Macosko "Rheology: Principles, Measurements and Applications", 1st Edition 1994, Wiley-VCH. C.W. Macosko "RIM Fundamentals of Reaction Injection Molding", 1st Edition 1989, Hanser Publishing References External links Chris Macosko, University of Minnesota Research Website - Chris Macosko Google Scholar - Chris W. Macosko Academic Tree - Chris Macosko Youtube: CEMS-UMN: Chris Macosko's "Last Lecture, March 1, 2018 Members of the United States National Academy of Engineering American chemical engineers Chemical engineers American materials scientists University of Minnesota faculty University of Minnesota alumni 1944 births Living people Minnesota CEMS
77360
https://en.wikipedia.org/wiki/GNOME%20Files
GNOME Files
GNOME Files, formerly and internally known as Nautilus, is the official file manager for the GNOME desktop. Nautilus was originally developed by Eazel with many luminaries from the tech world including Andy Hertzfeld (Apple), chief architect for Nautilus. The nautilus name was a play on words, evoking the shell of a nautilus to represent an operating system shell. Nautilus replaced Midnight Commander in GNOME 1.4 (2001) and has been the default file manager from version 2.0 onwards. Nautilus was the flagship product of the now-defunct Eazel Inc and was released under the terms of the GNU Lesser General Public License. It is free and open-source software. History Nautilus was originally developed by Eazel and Andy Hertzfeld (founder of Eazel and a former Apple engineer) in 1999. Nautilus was first released in 2001 and development has continued ever since. The following is a brief timeline of its development history: Version 1.0 was released on March 13, 2001, and incorporated into GNOME 1.4. Version 2.0 was a port to GTK+ 2.0. Version 2.2 included changes to make it more compliant with User Interface Guidelines. Version 2.4 switched the desktop folder to ~/Desktop (the ~ represents the user's "Home" folder) to be compliant with freedesktop.org standards. In the version included with GNOME 2.6, Nautilus switched to a spatial interface. Several Linux distributions have made "browser" mode the default. The "classic" interface is still available: By a filing cabinet shaped icon. By an option in the "Edit -> Preferences -> Behavior" menu in Nautilus. In a folder's context menu. By using the "--browser" switch when started by a command via a launcher or shell. GNOME 2.14 introduced a version of Nautilus with improved searching, integrated optional Beagle support and the ability to save searches as virtual folders. With the release of GNOME 2.22, Nautilus was ported to the newly introduced GVfs, the replacement virtual file system for the aging GnomeVFS. The 2.24 stable release of Nautilus adds some new features, mainly tabbed browsing and better tab completion. With GNOME 2.30, Nautilus reverted from a spatial interface to a browser navigational model by default. The 2.32 release introduced a dialog for handling conflicts when performing copy or move operations, transparency icon effect when cutting files into folder and enhanced the Wastebucket with Restore files. Besides, this is the last version that is based on GTK2 before the move to GNOME 3.0 with GTK3. Nautilus 2.x was forked to Caja, as well as MATE Desktop from Gnome 2.x after Gnome 3.0. Today both Mate and Caja are based on GTK3. GNOME 3.0 completely revamped the UX of Nautilus with focus on sidebar and icons. Additionally, the Connect to Server dialog is also enhanced. Nautilus was ported to GTK3. Version 3.4 added Undo functionality. Version 3.6 introduced a revamped UI design, symbolic sidebar icon, new search feature, removal of many features such as setting window background, emblems, split pane mode, spatial mode, scripts, compact view mode and tree view. Nautilus' application name was renamed to Files, Though it is still called Nautilus internally in some distributions. These major changes led to a lot of criticism, and various vendors such as Linux Mint decided to fork version 3.4. Version 3.8 included a new option to view files and folders as a tree, a new Connect to Server item in the sidebar and incremental loading of search results. Version 3.10 introduced a slightly revamped UI design in which titlebars and toolbars were merged into a single element called header bars. Version 3.18 introduced integration with Google Drive and GOA (gnome-online-accounts) settings. Features Bookmarks, window backgrounds, notes, and add-on scripts are all implemented, and the user has the choice between icon, list, or compact list views. In browser mode, Nautilus keeps a history of visited folders, similar to web browsers, permitting quick revisiting of folders. Nautilus can display previews of files in their icons, be they text files, images, sound or video files via thumbnailers such as Totem. Audio files are previewed (played back over GStreamer) when the pointer is hovering over them. In earlier versions, Nautilus included original vectorized icons designed by Susan Kare. File system abstraction GNOME Files relies on a file system abstraction layer (provided by GVfs) to browse local and remote file systems, including but not limited to FTP sites, Windows SMB shares, OBEX protocol (often implemented on cellphones), files transferred over shell protocol, HTTP and WebDAV and SFTP servers. Using the GIO library, Nautilus tracks modification of local files in real time, eliminating the need to refresh the display. GIO internally supports Gamin and FAM, Linux's inotify and Solaris' File Events Notification system. File indexing and file search framework GNOME Files relies on Tracker (formerly named "MetaTracker") to index file and is hence able to provide fast file search results. Batch renaming Batch renaming was introduced with GNOME Files version 3.22 (2016). Archive handling GNOME Files version 3.22 adds native, integrated file compression and decompression. By default, handling of archive files (e.g. .tar.gz) was handed off to File Roller (or another tool). Users now benefit from a progress bar, undo support, and an archive creation wizard. The new "extract on open" behavior, which automatically extracts an archive file by double clicking it, can be disabled in the preferences. MIME types MIME types (also called "media type" or "content type") are standardized by the IANA, then the freedesktop.org project takes care that the implementation works across all free software desktops. shared-mime-info is the provided library. At this time, at least GNOME, KDE, Xfce and ROX use this database. See also Comparison of file managers Dolphin (file manager) - KDE file manager Caja – a fork of Nautilus 2 for MATE Nemo – a fork of Nautilus 3 for Cinnamon References External links Introduction to spatial Nautilus 2001 software File managers that use GTK Free file managers Free software programmed in C GNOME Core Applications Software that uses Meson
78261
https://en.wikipedia.org/wiki/High-Level%20Data%20Link%20Control
High-Level Data Link Control
High-Level Data Link Control (HDLC) is a bit-oriented code-transparent synchronous data link layer protocol developed by the International Organization for Standardization (ISO). The standard for HDLC is ISO/IEC 13239:2002. HDLC provides both connection-oriented and connectionless service. HDLC can be used for point-to-multipoint connections via the original master-slave modes Normal Response Mode (NRM) and Asynchronous Response Mode (ARM), but they are now rarely used; it is now used almost exclusively to connect one device to another, using Asynchronous Balanced Mode (ABM). History HDLC is based on IBM's SDLC protocol, which is the layer 2 protocol for IBM's Systems Network Architecture (SNA). It was extended and standardized by the ITU as LAP (Link Access Procedure), while ANSI named their essentially identical version ADCCP. The HDLC specification does not specify the full semantics of the frame fields. This allows other fully compliant standards to be derived from it, and derivatives have since appeared in innumerable standards. It was adopted into the X.25 protocol stack as LAPB, into the V.42 protocol as LAPM, into the Frame Relay protocol stack as LAPF and into the ISDN protocol stack as LAPD. The original ISO standards for HDLC are the following: ISO 3309-1979 – Frame Structure ISO 4335-1979 – Elements of Procedure ISO 6159-1980 – Unbalanced Classes of Procedure ISO 6256-1981 – Balanced Classes of Procedure ISO/IEC 13239:2002, the current standard, replaced all of these specifications. HDLC was the inspiration for the IEEE 802.2 LLC protocol, and it is the basis for the framing mechanism used with the PPP on synchronous lines, as used by many servers to connect to a WAN, most commonly the Internet. A similar version is used as the control channel for E-carrier (E1) and SONET multichannel telephone lines. Cisco HDLC uses low-level HDLC framing techniques but adds a protocol field to the standard HDLC header. Framing HDLC frames can be transmitted over synchronous or asynchronous serial communication links. Those links have no mechanism to mark the beginning or end of a frame, so the beginning and end of each frame has to be identified. This is done by using a unique sequence of bits as a frame delimiter, or flag, and encoding the data to ensure that the flag sequence is never seen inside a frame. Each frame begins and ends with a frame delimiter. A frame delimiter at the end of a frame may also mark the start of the next frame. On both synchronous and asynchronous links, the flag sequence is binary "01111110", or hexadecimal 0x7E, but the details are quite different. Synchronous framing Because a flag sequence consists of six consecutive 1-bits, other data is coded to ensure that it never contains more than five 1-bits in a row. This is done by bit stuffing: any time that five consecutive 1-bits appear in the transmitted data, the data is paused and a 0-bit is transmitted. The receiving device knows that this is being done, and after seeing five 1-bits in a row, a following 0-bit is stripped out of the received data. If instead the sixth bit is 1, this is either a flag (if the seventh bit is 0), or an error (if the seventh bit is 1). In the latter case, the frame receive procedure is aborted, to be restarted when a flag is next seen. This bit-stuffing serves a second purpose, that of ensuring a sufficient number of signal transitions. On synchronous links, the data is NRZI encoded, so that a 0-bit is transmitted as a change in the signal on the line, and a 1-bit is sent as no change. Thus, each 0 bit provides an opportunity for a receiving modem to synchronize its clock via a phase-locked loop. If there are too many 1-bits in a row, the receiver can lose count. Bit-stuffing provides a minimum of one transition per six bit times during transmission of data, and one transition per seven bit times during transmission of a flag. When no frames are being transmitted on a simplex or full-duplex synchronous link, a frame delimiter is continuously transmitted on the link. This generates one of two continuous waveforms, depending on the initial state: The HDLC specification allows the 0-bit at the end of a frame delimiter to be shared with the start of the next frame delimiter, i.e. "011111101111110". Some hardware does not support this. For half-duplex or multi-drop communication, where several transmitters share a line, a receiver on the line will see continuous idling 1-bits in the inter-frame period when no transmitter is active. HDLC transmits bytes of data with the least significant bit first (not to be confused with little-endian order, which refers to byte ordering within a multi-byte field). Asynchronous framing When using asynchronous serial communication such as standard RS-232 serial ports, synchronous-style bit stuffing is inappropriate for several reasons: Bit stuffing is not needed to ensure an adequate number of transitions, as start and stop bits provide that, Because the data is NRZ encoded for transmission, rather than NRZI encoded, the encoded waveform is different, RS-232 sends bits in groups of 8, making adding single bits very awkward, and For the same reason, it is only necessary to specially code flag bytes; it is not necessary to worry about the bit pattern straddling multiple bytes. Instead asynchronous framing uses "control-octet transparency", also called "byte stuffing" or "octet stuffing". The frame boundary octet is 01111110, (0x7E in hexadecimal notation). A "control escape octet", has the value 0x7D (bit sequence '10111110', as RS-232 transmits least-significant bit first). If either of these two octets appears in the transmitted data, an escape octet is sent, followed by the original data octet with bit 5 inverted. For example, the byte 0x7E would be transmitted as 0x7D 0x5E ("10111110 01111010"). Other reserved octet values (such as XON or XOFF) can be escaped in the same way if necessary. The "abort sequence" 0x7D 0x7E ends a packet with an incomplete byte-stuff sequence, forcing the receiver to detect an error. This can be used to abort packet transmission with no chance the partial packet will be interpreted as valid by the receiver. Structure The contents of an HDLC frame are shown in the following table: Note that the end flag of one frame may be (but does not have to be) the beginning (start) flag of the next frame. Data is usually sent in multiples of 8 bits, but only some variants require this; others theoretically permit data alignments on other than 8-bit boundaries. The frame check sequence (FCS) is a 16-bit CRC-CCITT or a 32-bit CRC-32 computed over the Address, Control, and Information fields. It provides a means by which the receiver can detect errors that may have been induced during the transmission of the frame, such as lost bits, flipped bits, and extraneous bits. However, given that the algorithms used to calculate the FCS are such that the probability of certain types of transmission errors going undetected increases with the length of the data being checked for errors, the FCS can implicitly limit the practical size of the frame. If the receiver's calculation of the FCS does not match that of the sender's, indicating that the frame contains errors, the receiver can either send a negative acknowledge packet to the sender, or send nothing. After either receiving a negative acknowledge packet or timing out waiting for a positive acknowledge packet, the sender can retransmit the failed frame. The FCS was implemented because many early communication links had a relatively high bit error rate, and the FCS could readily be computed by simple, fast circuitry or software. More effective forward error correction schemes are now widely used by other protocols. Types of stations (computers) and data transfer modes Synchronous Data Link Control (SDLC) was originally designed to connect one computer with multiple peripherals via a multidrop bus. The original "normal response mode" is a master-slave mode where the computer (or primary terminal) gives each peripheral (secondary terminal) permission to speak in turn. Because all communication is either to or from the primary terminal, frames include only one address, that of the secondary terminal; the primary terminal is not assigned an address. There is a distinction between commands sent by the primary to a secondary, and responses sent by a secondary to the primary, but this is not reflected in the encoding; commands and responses are indistinguishable except for the difference in the direction in which they are transmitted. Normal response mode allows the secondary-to-primary link to be shared without contention, because it has the primary give the secondaries permission to transmit one at a time. It also allows operation over half-duplex communication links, as long as the primary is aware that it may not transmit when it has given permission to a secondary. Asynchronous response mode is an HDLC addition for use over full-duplex links. While retaining the primary/secondary distinction, it allows the secondary to transmit at any time. Thus, there must be some other mechanism to ensure that multiple secondaries do not try to transmit at the same time (or only one secondary). Asynchronous balanced mode adds the concept of a combined terminal which can act as both a primary and a secondary. Unfortunately, this mode of operation has some implementation subtleties. While the most common frames sent do not care whether they are in a command or response frame, some essential ones do (notably most unnumbered frames, and any frame with the P/F bit set), and the address field of a received frame must be examined to determine whether it contains a command (the address received is ours) or a response (the address received is that of the other terminal). This means that the address field is not optional, even on point-to-point links where it is not needed to disambiguate the peer being talked to. Some HDLC variants extend the address field to include both source and destination addresses, or an explicit command/response bit. HDLC operations and frame types Three fundamental types of HDLC frames may be distinguished: Information frames, or I-frames, transport user data from the network layer. They can also include flow and error control information piggybacked on data. Supervisory frames, or S-frames, are used for flow and error control whenever piggybacking is impossible or inappropriate, such as when a station does not have data to send. S-frames do not have information fields. Unnumbered frames, or U-frames, are used for various miscellaneous purposes, including link management. Some U-frames contain an information field, depending on the type. Control field The general format of the control field is: There are also extended (two-byte) forms of I and S frames. Again, the least significant bit (rightmost in this table) is sent first. P/F bit Poll/Final is a single bit with two names. It is called Poll when part of a command (set by the primary station to obtain a response from a secondary station), and Final when part of a response (set by the secondary station to indicate a response or the end of transmission). In all other cases, the bit is clear. The bit is used as a token that is passed back and forth between the stations. Only one token should exist at a time. The secondary only sends a Final when it has received a Poll from the primary. The primary only sends a Poll when it has received a Final back from the secondary, or after a timeout indicating that the bit has been lost. In NRM, possession of the poll token also grants the addressed secondary permission to transmit. The secondary sets the F-bit in its last response frame to give up permission to transmit. (It is equivalent to the word "Over" in radio voice procedure.) In ARM and ABM, the P bit forces a response. In these modes, the secondary need not wait for a poll to transmit, so the final bit may be included in the first response after the poll. If no response is received to a P bit in a reasonable period of time, the primary station times out and sends P again. The P/F bit is at the heart of the basic checkpoint retransmission scheme that is required to implement HDLC; all other variants (such as the REJ S-frame) are optional and only serve to increase efficiency. Whenever a station receives a P/F bit, it may assume that any frames that it sent before it last transmitted the P/F bit and not yet acknowledged will never arrive, and so should be retransmitted. When operating as a combined station, it is important to maintain the distinction between P and F bits, because there may be two checkpoint cycles operating simultaneously. A P bit arriving in a command from the remote station is not in response to our P bit; only an F bit arriving in a response is. N(R), the receive sequence number Both I and S frames contain a receive sequence number N(R). N(R) provides a positive acknowledgement for the receipt of I-frames from the other side of the link. Its value is always the first frame not yet received; it acknowledges that all frames with N(S) values up to N(R)−1 (modulo 8 or modulo 128) have been received and indicates the N(S) of the next frame it expects to receive. N(R) operates the same way whether it is part of a command or response. A combined station only has one sequence number space. N(S), the sequence number of the sent frame This is incremented for successive I-frames, modulo 8 or modulo 128. Depending on the number of bits in the sequence number, up to 7 or 127 I-frames may be awaiting acknowledgment at any time. I-Frames (user data) Information frames, or I-frames, transport user data from the network layer. In addition they also include flow and error control information piggybacked on data. The sub-fields in the control field define these functions. The least significant bit (first transmitted) defines the frame type. 0 means an I-frame. Except for the interpretation of the P/F field, there is no difference between a command I frame and a response I frame; when P/F is 0, the two forms are exactly equivalent. S-frames (control) Supervisory Frames, or 'S-frames', are used for flow and error control whenever piggybacking is impossible or inappropriate, such as when a station does not have data to send. S-frames in HDLC do not have information fields, although some HDLC-derived protocols use information fields for "multi-selective reject". The S-frame control field includes a leading "10" indicating that it is an S-frame. This is followed by a 2-bit type, a poll/final bit, and a 3-bit sequence number. (Or a 4-bit padding field followed by a 7-bit sequence number.) The first (least significant) 2 bits mean it is an S-frame. All S frames include a P/F bit and a receive sequence number as described above. Except for the interpretation of the P/F field, there is no difference between a command S frame and a response S frame; when P/F is 0, the two forms are exactly equivalent. Receive Ready (RR) Bit value = 00 (0x00 to match above table type field bit order) Indicate that the sender is ready to receive more data (cancels the effect of a previous RNR). Send this packet if you need to send a packet but have no I frame to send. A primary station can send this with the P-bit set to solicit data from a secondary station. A secondary terminal can use this with the F-bit set to respond to a poll if it has no data to send. Receive Not Ready (RNR) Bit value = 10 (0x04 to match above table type field bit order) Acknowledge some packets but request no more be sent until further notice. Can be used like RR with P bit set to solicit the status of a secondary station Can be used like RR with F bit set to respond to a poll if the station is busy. Reject (REJ) Bit value = 01 (0x08 to match above table type field bit order) Requests immediate retransmission starting with N(R). Sent in response to an observed sequence number gap; e.g. after seeing I1/I2/I3/I5, send REJ4. Optional to generate; a working implementation may use only RR. Selective Reject (SREJ) Bit value = 11 (0x0c to match above table type field bit order) Requests retransmission of only the frame N(R). Not supported by all HDLC variants. Optional to generate; a working implementation may use only RR, or only RR and REJ. U-Frames Unnumbered frames, or U-frames, are primarily used for link management, although a few are used to transfer user data. They exchange session management and control information between connected devices, and some U-frames contain an information field, used for system management information or user data. The first 2 bits (11) mean it is a U-frame. The five type bits (2 before P/F bit and 3 bit after P/F bit) can create 32 different types of U-frame. In a few cases, the same encoding is used for different things as a command and a response. Mode setting The various modes are described in . Briefly, there are two non-operational modes (initialization mode and disconnected mode) and three operational modes (normal response, asynchronous response, and asynchronous balanced modes) with 3-bit or 7-bit (extended) sequence numbers. Disconnected mode (DM) response When the secondary is disconnected (the default state on power-up), it sends this generic response to any poll (command frame with the poll flag set) except an acceptable mode setting command. It may alternatively give a FRMR response to an unacceptable mode set command. Unnumbered acknowledge (UA) response This is the secondary's response to an acceptable mode set command, indicating that it is now in the requested mode. Set ... mode (SNRM, SARM, SABM) command Place the secondary in the specified mode, with 3-bit sequence numbers (1-byte control field). The secondary acknowledges with UA. If the secondary does not implement the mode, it responds with DM or FRMR. Set ... mode extended (SNRME, SARME, SABME) command Place the secondary in the specified mode, with 7-bit sequence numbers (2-byte control field). Set mode (SM) command Generic mode set, new in ISO/IEC 13239, using an information field to select parameters. ISO/IEC 13239 added many additional options to HDLC, including 15- and 31-bit sequence numbers, which can only be selected with this command. Disconnect (DISC) command This command causes the secondary to acknowledge with UA and disconnect (enter disconnected mode). Any unacknowledged frames are lost. Request disconnect (RD) response This response requests the primary to send a DISC command. The primary should do so promptly, but may delay long enough to ensure all pending frames are acknowledged. Set initialization mode (SIM) command This rarely-implemented command is used to perform some secondary-specific initialization, such as downloading firmware. What happens in initialization mode is not otherwise specified in the HDLC standard. Request initialization mode (RIM) response This requests the primary to send SIM and initialize the secondary. It sent in lieu of DM if the secondary requires initialization. Information transfer These frames may be used as part of normal information transfer. Unnumbered information (UI) This frame (command or response) communicates user data, but without acknowledgement or retransmission in case of error. UI with header check (UIH) This frame (command or response), a ISO/IEC 13239 addition and rarely used, is like UI but also excludes CRC protection. Only a configurable-length prefix ("header") of the frame is covered by the CRC polynomial; errors in the rest of the frame are not detected. Unnumbered poll (UP) command This command solicits a response from the secondary. With the poll bit set, it acts like any other poll frame, without the acknowledgement that must be included in I or S frame. With the poll bit clear, it has a special meaning in normal response mode: the secondary may respond, even though it has not received the poll bit. This is rarely used in HDLC, but was used in the original IBM SDLC as a substitute for the lack of asynchronous response mode; where the communication channel could handle simultaneous responses, the primary would periodically send UP to the broadcast address to collect any pending responses. Error Recovery Frame reject (FRMR) response The FRMR response contains a description of the unacceptable frame, in a standardized format. The first 1 or 2 bytes are a copy of the rejected control field, the next 1 or 2 contain the secondary's current send and receive sequence numbers (and a flag indicating that the frame was a response, applicable only in balanced mode), and the following 4 or 5 bits are error flags indicating the reason for the rejection. The secondary repeats the same FRMR response to every poll until the error is cleared by a mode set command or RSET. The error flags are: W: the frame type (control field) is not understood or not implemented. X: the frame type is not understood with a non-empty information field, but one was present. Y: the frame included an information field that is larger than the secondary can accept. Z: the frame included an invalid receive sequence number N(R), one which is not between the previously received value and the highest sequence number transmitted. (This error cannot be cleared by receiving RSET, but can be cleared by sending RSET.) V: the frame included an invalid send sequence number N(S), greater than the last number acknowledged plus the transmit window size. This error is only possible if a transmit window size smaller than the maximum has been negotiated. The error flags are normally padded with 0 bits to an 8-bit boundary, but HDLC permits frames which are not a multiple of a byte long. Reset (RSET) command The RSET command causes a secondary to reset its receive sequence number so the next expected frame is sequence number 0. This is a possible alternative to sending a new mode set command, which resets both sequence numbers. It is acknowledged with UA, like a mode set command. Peer discovery Exchange identification (XID) An XID command includes an information field specifying the primary's capabilities; the secondary responds with an XID response specifying its capabilities. This is normally done before sending a mode set command. Systems Network Architecture defined one format for the information field, in which the most significant bit of the first byte is clear (0), but HDLC implementations normally implement the variant defined in ISO 8885, which has the most significant bit of the first byte set (1). TEST A TEST command is simply a ping command for debugging purposes. The payload of the TEST command is returned in the TEST response. Defined in other standards There are several U frames which are not part of HDLC, but defined in other related standards. Nonreserved (NR0, NR1, NR2, NR3) The "nonreserved" commands and responses are guaranteed by the HDLC standard to be available for other uses. Ack connectionless (AC0, AC1) These are defined in the IEEE 802.2 logical link control standard. Configure (CFGR) This command was defined in SDLC for debugging. It had a 1-byte payload which specified a non-standard test mode for the secondary. Even numbers disabled the mode, while odd numbers enabled it. A payload of 0 disabled all test modes. The secondary normally acknowledged a configure command by echoing it in response. Beacon (BCN) response This response was defined in SDLC to indicate a communications failure. A secondary which received no frames at all for a long time would begin sending a stream of beacon responses, allowing a unidirectional fault to be located. Note that ISO/IEC 13239 assigns UIH the same encoding as BCN. Link configurations Link configurations can be categorized as being either: Unbalanced, which consists of one primary terminal, and one or more secondary terminals. Balanced, which consists of two peer terminals. The three link configurations are: Normal Response Mode (NRM) is an unbalanced configuration in which only the primary terminal may initiate data transfer. The secondary terminals transmit data only in response to commands from the primary terminal. The primary terminal polls each secondary terminal to give it an opportunity to transmit any data it has. Asynchronous Response Mode (ARM) is an unbalanced configuration in which secondary terminals may transmit without permission from the primary terminal. However, there is still a distinguished primary terminal which retains responsibility for line initialization, error recovery, and logical disconnect. Asynchronous Balanced Mode (ABM) is a balanced configuration in which either station may initialize, supervise, recover from errors, and send frames at any time. There is no master/slave relationship. The DTE (data terminal equipment) and DCE (data circuit-terminating equipment) are treated as equals. The initiator for Asynchronous Balanced Mode sends an SABM. An additional link configuration is Disconnected mode. This is the mode that a secondary station is in before it is initialized by the primary, or when it is explicitly disconnected. In this mode, the secondary responds to almost every frame other than a mode set command with a "Disconnected mode" response. The purpose of this mode is to allow the primary to reliably detect a secondary being powered off or otherwise reset. HDLC Command and response repertoire The minimal set required for operation are: Commands: I, RR, RNR, DISC, and one of SNRM, SARM or SABM Responses: I, RR, RNR, UA, DM, FRMR Basic operations Initialization can be requested by either side. When the primary sends one of the six mode-set commands, it: Signals the other side that initialization is requested Specifies the mode, NRM, ABM, ARM Specifies whether 3 or 7 bit sequence numbers are in use. The HDLC module on the other end transmits (UA) frame when the request is accepted. If the request is rejected it sends (DM) disconnect mode frame. Functional extensions (options) For Switched Circuits Commands: ADD – XID Responses: ADD – XID, RD For 2-way Simultaneous commands & responses are ADD – REJ For Single Frame Retransmission commands & responses: ADD – SREJ For Information Commands & Responses: ADD – Ul For Initialization Commands: ADD – SIM Responses: ADD – RIM For Group Polling Commands: ADD – UP Extended Addressing Delete Response I Frames Delete Command I Frames Extended Numbering For Mode Reset (ABM only) Commands are: ADD – RSET Data Link Test Commands & Responses are: ADD – TEST Request Disconnect. Responses are ADD – RD 32-bit FCS HDLC command and response repertoire Unnumbered frames Unnumbered frames are identified by the low two bits being 1. With the P/F flag, that leaves 5 bits as a frame type. Even though fewer than 32 values are in use, some types have different meanings depending on the direction they are sent: as a command or as a response. The relationship between the DISC (disconnect) command and the RD (request disconnect) response seems clear enough, but the reason for making SARM command numerically equal to the DM response is obscure. * ISO/IEC 13239 addition † Not part of HDLC The UI, UIH, XID, TEST frames contain a payload, and can be used as both commands and responses. The SM command and FRMR response also contain a payload. A UI frame contains user information, but unlike an I frame it is neither acknowledged nor retransmitted if lost. A UIH frame (an ISO/IEC 13239 addition) is like a UI frame, but additionally applies the frame check sequence only to a specified-length prefix of the frame; transmission errors after this prefix are not detected. The XID frame is used to exchange terminal capabilities. Systems Network Architecture defined one format, but the variant defined in ISO 8885 is more commonly used. A primary advertises its capabilities with an XID command, and a secondary returns its own capabilities in an XID response. The TEST frame is simply a ping command for debugging purposes. The payload of the TEST command is returned in the TEST response. The SM command (an ISO/IEC 13239 addition) is a generic "set mode" command which includes an information field (in the same ISO 8885 format as XID) specifying parameters. This allows parameter values (like 15- and 31-bit sequence numbers) and parameters like window sizes and maximum frame sizes not expressible by the standard six mode-set commands to be negotiated. The FRMR response contains a description of the unacceptable frame, in a standardized format. The first 1 or 2 bytes are a copy of the rejected control field, the next 1 or 2 contain the secondary's current send and receive sequence numbers, and the following 4 or 5 bits are error flags indicating the reason for the rejection. See also Point-to-Point Protocol Serial line IP Notes References External links Data Communication Lectures of Manfred Lindner – Part HDLC HDLC packet format and other information ISO 3309:1984 Information Processing Systems—Data Communication—High Level Data Link Control Procedures—Frame Structure (archived) ISO 4335:1984 Data Communication—High Level Data Link Control Procedures—Consolidation of Elements of Procedures (archived) ISO/IEC 13239:2002 Link protocols Logical link control Telecommunications standards Networking standards
14513000
https://en.wikipedia.org/wiki/People%27s%20Computer%20Company
People's Computer Company
People's Computer Company (PCC) was an organization, a newsletter (the People's Computer Company Newsletter) and, later, a quasiperiodical called the Dragonsmoke. PCC was founded and produced by Dennis Allison, Bob Albrecht and George Firedrake in Menlo Park, California in the early 1970s. The first newsletter, published in October 1972, announced itself with the following introduction: "Computers are mostly used against people instead of for people; used to control people instead of to free them; Time to change all that - we need a... Peoples Computer Company." It was published bimonthly. The name was chosen in reference to Janis Joplin’s rock group Big Brother and the Holding Company. The newsletter ceased publication in 1981. History PCC was one of the first organizations to recognize the potential of Tiny BASIC in the nascent field of personal computing when it published that language's design specification in their newsletter. This ultimately led to the design of an interpreter that was published in a publication, which they named Dr. Dobb's Journal of Tiny BASIC Calisthenics and Orthodontia, dedicated to Tiny BASIC. The newsletter's title was changed to Dr. Dobb's Journal of Computer Calisthenics & Orthodontia for the second issue; the popular reaction to it eventually led to the long-running computer magazine Dr. Dobb's Journal (DDJ) which continued publication until 2009. PCC was among the first organizations to recognize and actively advocate playing as a legitimate way of learning. It published arguably the first best-seller in microcomputer literature, My Computer Likes Me When I Speak BASIC and What to Do After You Hit Return. The company was an early proponent of software without copyright, and published much of it in the above books, in DDJ and in another periodical. That magazine originally shared the company's name but it evolved and was later renamed Recreational Computing. It focused on publishing code listings, mostly for games, that users could hand type into their early-model (and some homebrew) personal computers. Because the code was without copyright, authors were free to study it, adapt, rewrite and build upon it. The same was true of the more systems-oriented code published in DDJ. This no-copyright practice was a significant boost to the growing body of microcomputer software and applications, and to the general base of knowledge and developing best practices in the young industry. PCC also fostered the activities of its child organization, ComputerTown USA! That formalized PCC's long-standing activism around general computer literacy. At a time when many computers still were kept in clean rooms, PCC was taking them to libraries, grade schools and elder communities. Their activities encouraged hands-on exploration and just trying things. The Logo programming language and turtle graphics gave some users their first experience of controlling something on a computer display. Computer phobia was commonly perceived by PCC staff as a barrier to learning in a significant number of users, even in a large majority of some populations during those early years. Apple Computer's Community Affairs department used ComputerTown USA! to develop curriculum and to conduct intensive trainings for the non-profit recipients of computer hardware and software grants from Apple. As one of its core philosophical contributions, People's Computer Company recognized in personal computing a great potential for individual empowerment and social improvement. It saw that PCs could bring the same advantages to those hampered by race, class and circumstance as to those with more advantages. It believed a digital commons could lead to more intermingling of individuals from diverse social groups. It supported early models of networking personal computers using telephone lines. It could, in hindsight, be regarded as among the first contributors to a form of what is known now as network neutrality and a fundamentally non-commercial, class-free internet. The history of PCC and its role in the evolution of the personal computer was described in Steven Levy's book, Hackers: Heroes of the Computer Revolution. In Levy's book, some of the values and ethics of PCC's founders are examined, particularly the ethics common among members of the hacker community. See also Hacker ethic Community Memory Microcomputer revolution References External links Site that is the home of the history of the Peoples Computer Company The PCC alumni pages Digital Collection of People's Computer Company publications Bimonthly magazines published in the United States Computer clubs Defunct computer magazines published in the United States Magazines established in 1972 Magazines disestablished in 1981 Magazines published in California Menlo Park, California
19689118
https://en.wikipedia.org/wiki/Gary%20McGraw
Gary McGraw
Gary McGraw is an American computer scientist, author, and researcher. Education McGraw holds a dual PhD in Cognitive Science and Computer Science from Indiana University and a BA in Philosophy from the University of Virginia. His doctoral dissertation is titled "Letter Spirit: Emergent High-Level Perception of Letters Using Fluid Concepts." Career McGraw was the Vice President of Security Technology at Synopsys. Before Cigital was acquired by Synopsys, he was Chief Technical Officer at Cigital. He produced the Silver Bullet Security Podcast for IEEE Security & Privacy magazine (syndicated by informIT). Gary McGraw serves on the Dean's Advisory Council for the School of Informatics of Indiana University. He also serves on the advisory boards of several companies, including Dasient (acquired by Twitter), Fortify Software (acquired by Hewlett-Packard), Max Financial, Invotas, Wall+Main, Invincea (acquired by Sophos), and Raven White. In the past, Gary McGraw has served on the IEEE Computer Society Board of Governors. Books Gary is an author of many books and over 100 peer-reviewed publications on IT security. Software Security: Building Security In, Exploiting Software: How to Break Code (with Greg Hoglund), Building Secure Software: How to Avoid Security Problems the Right Way (with John Viega), Java Security (with Edward Felten), Exploiting Online Games: Cheating Massively Distributed Systems (with Greg Hoglund), Software Security Engineering: A Guide for Project Managers (with Julia H. Allen, Sean J. Barnum, Robert J. Ellison, and Nancy R. Mead) Software Fault Injection (with Jeffrey M. Voas) Securing Java: Getting Down to Business with Mobile Code (with Edward Felten), Notes References Ben Rothke. "Software Security: Building Security In", Security Management magazine Radu State. Review of "Software Security: Building Security In by Gary McGraw", ACM Queue 4(7):44 (2006) "Software Security : Building Security In", Palizine, Issue #18 February 2006 Robert Bruen. "Software Security. Building Security In", Cipher (IEEE magazine), Jan 5, 2006 Alen Prodan. "Exploiting Software: How to Break Code", Help Net Security, 21 July 2004 A. Mariën. Review of "Exploiting Software: How to Break Code by Greg Hoglund and Gary McGraw", ACM Queue, 3(4):60 (2005) Robert Bruen. "Exploiting Software. How to Break Code", Cipher (IEEE magazine), January 13, 2004 Aleksandar Stancin. "Building Secure Software: How to Avoid Security Problems the Right Way", Help Net Security Robert Bruen. "Building Secure Software. How to Avoid Security Problems the Right Way", Cipher (IEEE magazine), January 9, 2002 Diomidis Spinellis. "Book review: Building Secure Software: how to Avoid Security Problems the Right Way", ACM Computing Reviews, 43(4):103–104, April 2002. External links Gary McGraw's personal home page 1966 births Living people American technology writers Writers about computer security Indiana University alumni University of Virginia alumni
238775
https://en.wikipedia.org/wiki/Newline
Newline
Newline (frequently called line ending, end of line (EOL), next line (NEL) or line break) is a control character or sequence of control characters in a character encoding specification (e.g., ASCII, EBCDIC) that is used to signify the end of a line of text and the start of a new one. History In the mid-1800s, long before the advent of teleprinters and teletype machines, Morse code operators or telegraphists invented and used Morse code prosigns to encode white space text formatting in formal written text messages. In particular the Morse prosign (mnemonic reak ext) represented by the concatenation of literal textual Morse codes "B" and "T" characters sent without the normal inter-character spacing is used in Morse code to encode and indicate a new line or new section in a formal text message. Later, in the age of modern teleprinters, standardized character set control codes were developed to aid in white space text formatting. ASCII was developed simultaneously by the International Organization for Standardization (ISO) and the American Standards Association (ASA), the latter being the predecessor organization to American National Standards Institute (ANSI). During the period of 1963 to 1968, the ISO draft standards supported the use of either + or alone as a newline, while the ASA drafts supported only +. The sequence + was commonly used on many early computer systems that had adopted Teletype machines—typically a Teletype Model 33 ASR—as a console device, because this sequence was required to position those printers at the start of a new line. The separation of newline into two functions concealed the fact that the print head could not return from the far right to the beginning of the next line in time to print the next character. Any character printed after a would often print as a smudge in the middle of the page while the print head was still moving the carriage back to the first position. "The solution was to make the newline two characters: to move the carriage to column one, and to move the paper up." In fact, it was often necessary to send extra characters—extraneous CRs or NULs—which are ignored but give the print head time to move to the left margin. Many early video displays also required multiple character times to scroll the display. On such systems, applications had to talk directly to the Teletype machine and follow its conventions since the concept of device drivers hiding such hardware details from the application was not yet well developed. Therefore, text was routinely composed to satisfy the needs of Teletype machines. Most minicomputer systems from DEC used this convention. CP/M also used it in order to print on the same terminals that minicomputers used. From there MS-DOS (1981) adopted CP/M's + in order to be compatible, and this convention was inherited by Microsoft's later Windows operating system. The Multics operating system began development in 1964 and used alone as its newline. Multics used a device driver to translate this character to whatever sequence a printer needed (including extra padding characters), and the single byte was more convenient for programming. What seems like a more obvious choice——was not used, as provided the useful function of overprinting one line with another to create boldface, underscore and strikethrough effects. Perhaps more importantly, the use of alone as a line terminator had already been incorporated into drafts of the eventual ISO/IEC 646 standard. Unix followed the Multics practice, and later Unix-like systems followed Unix. This created conflicts between Windows and Unix-like operating systems, whereby files composed on one operating system cannot be properly formatted or interpreted by another operating system (for example a UNIX shell script written in a Windows text editor like Notepad). Representation The concepts of carriage return (CR) and line feed (LF) are closely associated and can be considered either separately or together. In the physical media of typewriters and printers, two axes of motion, "down" and "across", are needed to create a new line on the page. Although the design of a machine (typewriter or printer) must consider them separately, the abstract logic of software can combine them together as one event. This is why a newline in character encoding can be defined as and combined into one (commonly called or ). Some character sets provide a separate newline character code. EBCDIC, for example, provides an character code in addition to the and codes. Unicode, in addition to providing the ASCII and control codes, also provides a "next line" () control code, as well as control codes for "line separator" and "paragraph separator" markers. EBCDIC systems—mainly IBM mainframe systems, including z/OS (OS/390) and IBM i (OS/400)—use (New Line, ) as the character combining the functions of line feed and carriage return. The equivalent Unicode character () is called (Next Line). EBCDIC also has control characters called and , but the numerical value of () differs from the one used by ASCII (). Additionally, some EBCDIC variants also use but assign a different numeric code to the character. However, those operating systems use a record-based file system, which stores text files as one record per line. In most file formats, no line terminators are actually stored. Operating systems for the CDC 6000 series defined a newline as two or more zero-valued six-bit characters at the end of a 60-bit word. Some configurations also defined a zero-valued character as a colon character, with the result that multiple colons could be interpreted as a newline depending on position. RSX-11 and OpenVMS also use a record-based file system, which stores text files as one record per line. In most file formats, no line terminators are actually stored, but the Record Management Services facility can transparently add a terminator to each line when it is retrieved by an application. The records themselves could contain the same line terminator characters, which could either be considered a feature or a nuisance depending on the application. RMS not only stored records, but also stored metadata about the record separators in different bits for the file to complicate matters even more (since files could have fixed length records, records that were prefixed by a count or records that were terminated by a specific character). The bits were not generic, so while they could specify that or or even was the line terminator, it could not substitute some other code. Fixed line length was used by some early mainframe operating systems. In such a system, an implicit end-of-line was assumed every 72 or 80 characters, for example. No newline character was stored. If a file was imported from the outside world, lines shorter than the line length had to be padded with spaces, while lines longer than the line length had to be truncated. This mimicked the use of punched cards, on which each line was stored on a separate card, usually with 80 columns on each card, often with sequence numbers in columns 73–80. Many of these systems added a carriage control character to the start of the next record; this could indicate whether the next record was a continuation of the line started by the previous record, or a new line, or should overprint the previous line (similar to a ). Often this was a normal printing character such as that thus could not be used as the first character in a line. Some early line printers interpreted these characters directly in the records sent to them. Unicode The Unicode standard defines a number of characters that conforming applications should recognize as line terminators: : Line Feed, : Vertical Tab, : Form Feed, : Carriage Return, +: () followed by () : Next Line, : Line Separator, : Paragraph Separator, This may seem overly complicated compared to an approach such as converting all line terminators to a single character, for example . However, Unicode was designed to preserve all information when converting a text file from any existing encoding to Unicode and back. Therefore, Unicode should contain characters included in existing encodings. For example: is part of EBCDIC, which uses code ; it is normally mapped to Unicode , , which is a control character in the C1 control set. As such, it is defined by ECMA 48, and recognized by encodings compliant with ISO/IEC 2022 (which is equivalent to ECMA 35). C1 control set is also compatible with ISO-8859-1. The approach taken in the Unicode standard allows round-trip transformation to be information-preserving while still enabling applications to recognize all possible types of line terminators. Recognizing and using the newline codes greater than (, and ) is not often done. They are multiple bytes in UTF-8, and the code for has been used as the ellipsis () character in Windows-1252. For instance: ECMAScript accepts and as line-breaks, but considers () whitespace instead of a line-break. Windows 10 does not treat any of , , or as line-breaks in its default text editor, Notepad. gedit, the default text editor of the GNOME desktop environment, treats and as newlines but does not for . JSON allows and characters within strings, while ECMAScript prior to treated them as newlines, and therefore illegal syntax. YAML no longer recognizes them as special as of version 1.2, in order to be compatible with JSON. Note well that the Unicode special characters (, ), (, ), (, ) and (, ) are glyphs intended for presenting a user-visible character to the reader of the document, and are thus not recognized themselves as a newline. In programming languages To facilitate the creation of portable programs, programming languages provide some abstractions to deal with the different types of newline sequences used in different environments. The C programming language provides the escape sequences (newline) and (carriage return). However, these are not required to be equivalent to the ASCII and control characters. The C standard only guarantees two things: Each of these escape sequences maps to a unique implementation-defined number that can be stored in a single value. When writing to a file, device node, or socket/fifo in text mode, is transparently translated to the native newline sequence used by the system, which may be longer than one character. When reading in text mode, the native newline sequence is translated back to . In binary mode, no translation is performed, and the internal representation produced by is output directly. On Unix platforms, where C originated, the native newline sequence is ASCII (), so was simply defined to be that value. With the internal and external representation being identical, the translation performed in text mode is a no-op, and Unix has no notion of text mode or binary mode. This has caused many programmers who developed their software on Unix systems simply to ignore the distinction completely, resulting in code that is not portable to different platforms. The C library function is best avoided in binary mode because any file not written with the Unix newline convention will be misread. Also, in text mode, any file not written with the system's native newline sequence (such as a file created on a Unix system, then copied to a Windows system) will be misread as well. Another common problem is the use of when communicating using an Internet protocol that mandates the use of ASCII + for ending lines. Writing to a text mode stream works correctly on Windows systems, but produces only on Unix, and something completely different on more exotic systems. Using in binary mode is slightly better. Many languages, such as C++, Perl, and Haskell provide the same interpretation of as C. C++ has an alternative I/O model where the manipulator can be used to output a newline (and flushes the stream buffer). Java, PHP, and Python provide the sequence (for ASCII +). In contrast to C, these are guaranteed to represent the values and , respectively. The Java I/O libraries do not transparently translate these into platform-dependent newline sequences on input or output. Instead, they provide functions for writing a full line that automatically add the native newline sequence, and functions for reading lines that accept any of , , or + as a line terminator (see ). The method can be used to retrieve the underlying line separator. Example: String eol = System.lineSeparator(); String lineColor = "Color: Red" + eol; Python permits "Universal Newline Support" when opening a file for reading, when importing modules, and when executing a file. Some languages have created special variables, constants, and subroutines to facilitate newlines during program execution. In some languages such as PHP and Perl, double quotes are required to perform escape substitution for all escape sequences, including and . In PHP, to avoid portability problems, newline sequences should be issued using the PHP_EOL constant. Example in C#: string eol = Environment.NewLine; string lineColor = "Color: Red" + eol; string eol2 = "\n"; string lineColor2 = "Color: Blue" + eol2; Issues with different newline formats The different newline conventions cause text files that have been transferred between systems of different types to be displayed incorrectly. Text in files created with programs which are common on Unix-like or classic Mac OS, appear as a single long line on most programs common to MS-DOS and Microsoft Windows because these do not display a single or a single as a line break. Conversely, when viewing a file originating from a Windows computer on a Unix-like system, the extra may be displayed as a second line break, as , or as at the end of each line. Furthermore, programs other than text editors may not accept a file, e.g. some configuration file, encoded using the foreign newline convention, as a valid file. The problem can be hard to spot because some programs handle the foreign newlines properly while others do not. For example, a compiler may fail with obscure syntax errors even though the source file looks correct when displayed on the console or in an editor. Modern text editors generally recognize all flavours of + newlines and allow users to convert between the different standards. Web browsers are usually also capable of displaying text files and websites which use different types of newlines. Even if a program supports different newline conventions, these features are often not sufficiently labeled, described, or documented. Typically a menu or combo-box enumerating different newline conventions will be displayed to users without an indication if the selection will re-interpret, temporarily convert, or permanently convert the newlines. Some programs will implicitly convert on open, copy, paste, or save—often inconsistently. Most textual Internet protocols (including HTTP, SMTP, FTP, IRC, and many others) mandate the use of ASCII + (, ) on the protocol level, but recommend that tolerant applications recognize lone (, ) as well. Despite the dictated standard, many applications erroneously use the C newline escape sequence () instead of the correct combination of carriage return escape and newline escape sequences (+) (see section Newline in programming languages above). This accidental use of the wrong escape sequences leads to problems when trying to communicate with systems adhering to the stricter interpretation of the standards instead of the suggested tolerant interpretation. One such intolerant system is the qmail mail transfer agent that actively refuses to accept messages from systems that send bare instead of the required +. The standard Internet Message Format for email states: "CR and LF MUST only occur together as CRLF; they MUST NOT appear independently in the body". The File Transfer Protocol can automatically convert newlines in files being transferred between systems with different newline representations when the transfer is done in "ASCII mode". However, transferring binary files in this mode usually has disastrous results: any occurrence of the newline byte sequence—which does not have line terminator semantics in this context, but is just part of a normal sequence of bytes—will be translated to whatever newline representation the other system uses, effectively corrupting the file. FTP clients often employ some heuristics (for example, inspection of filename extensions) to automatically select either binary or ASCII mode, but in the end it is up to users to make sure their files are transferred in the correct mode. If there is any doubt as to the correct mode, binary mode should be used, as then no files will be altered by FTP, though they may display incorrectly. Conversion between newline formats Text editors are often used for converting a text file between different newline formats; most modern editors can read and write files using at least the different ASCII / conventions. For example, the editor Vim can make a file compatible with the Windows Notepad text editor. Within vim :set fileformat=dos :wq Editors can be unsuitable for converting larger files or bulk conversion of many files. For larger files (on Windows NT/2000/XP) the following command is often used: D:\>TYPE unix_file | FIND /V "" > dos_file Special purpose programs to convert files between different newline conventions include , and , and , and . The command is available on virtually every Unix-like system and can be used to perform arbitrary replacement operations on single characters. A DOS/Windows text file can be converted to Unix format by simply removing all ASCII characters with $ tr -d '\r' < inputfile > outputfile or, if the text has only newlines, by converting all newlines to with $ tr '\r' '\n' < inputfile > outputfile The same tasks are sometimes performed with awk, sed, or in Perl if the platform has a Perl interpreter: $ awk '{sub("$","\r\n"); printf("%s",$0);}' inputfile > outputfile # UNIX to DOS (adding CRs on Linux and BSD based OS that haven't GNU extensions) $ awk '{gsub("\r",""); print;}' inputfile > outputfile # DOS to UNIX (removing CRs on Linux and BSD based OS that haven't GNU extensions) $ sed -e 's/$/\r/' inputfile > outputfile # UNIX to DOS (adding CRs on Linux based OS that use GNU extensions) $ sed -e 's/\r$//' inputfile > outputfile # DOS to UNIX (removing CRs on Linux based OS that use GNU extensions) $ perl -pe 's/\r?\n|\r/\r\n/g' inputfile > outputfile # Convert to DOS $ perl -pe 's/\r?\n|\r/\n/g' inputfile > outputfile # Convert to UNIX $ perl -pe 's/\r?\n|\r/\r/g' inputfile > outputfile # Convert to old Mac The command can identify the type of line endings: $ file myfile.txt myfile.txt: ASCII English text, with CRLF line terminators The Unix egrep (extended grep) command can be used to print filenames of Unix or DOS files (assuming Unix and DOS-style files only, no classic Mac OS-style files): $ egrep -L '\r\n' myfile.txt # show UNIX style file (LF terminated) $ egrep -l '\r\n' myfile.txt # show DOS style file (CRLF terminated) Other tools permit the user to visualise the EOL characters: $ od -a myfile.txt $ cat -e myfile.txt $ cat -v myfile.txt $ hexdump -c myfile.txt Interpretation Two ways to view newlines, both of which are self-consistent, are that newlines either separate lines or that they terminate lines. If a newline is considered a separator, there will be no newline after the last line of a file. Some programs have problems processing the last line of a file if it is not terminated by a newline. On the other hand, programs that expect newline to be used as a separator will interpret a final newline as starting a new (empty) line. Conversely, if a newline is considered a terminator, all text lines including the last are expected to be terminated by a newline. If the final character sequence in a text file is not a newline, the final line of the file may be considered to be an improper or incomplete text line, or the file may be considered to be improperly truncated. In text intended primarily to be read by humans using software which implements the word wrap feature, a newline character typically only needs to be stored if a line break is required independent of whether the next word would fit on the same line, such as between paragraphs and in vertical lists. Therefore, in the logic of word processing and most text editors, newline is used as a paragraph break and is known as a "hard return", in contrast to "soft returns" which are dynamically created to implement word wrapping and are changeable with each display instance. In many applications a separate control character called "manual line break" exists for forcing line breaks inside a single paragraph. The glyph for the control character for a hard return is usually a pilcrow (¶), and for the manual line break is usually a carriage return arrow (↵). Reverse and partial line feeds , (U+008D REVERSE LINE FEED, ISO/IEC 6429 8D, decimal 141) is used to move the printing position back one line (by reverse feeding the paper, or by moving a display cursor up one line) so that other characters may be printed over existing text. This may be done to make them bolder, or to add underlines, strike-throughs or other characters such as diacritics. Similarly, (U+008B PARTIAL LINE FORWARD, decimal 139) and (U+008C PARTIAL LINE BACKWARD, decimal 140) can be used to advance or reverse the text printing position by some fraction of the vertical line spacing (typically, half). These can be used in combination for subscripts (by advancing and then reversing) and superscripts (by reversing and then advancing), and may also be useful for printing diacritics. See also ASA carriage control characters C0 and C1 control codes End-of-file Line starve Page break Carriage return Enter key References External links The Unicode reference, see paragraph 5.8 in Chapter 5 of the Unicode 4.0 standard (PDF) The End of Line Puzzle "The End-of-Line Story" Control characters Whitespace
56448459
https://en.wikipedia.org/wiki/Luminar%20%28software%29
Luminar (software)
Luminar is a universal photo editing software application developed by Skylum (formerly Macphun) available for Windows and macOS. Luminar works as a standalone application and as a plugin for Adobe and Apple products. It integrates with Adobe Photoshop, Adobe Photoshop Elements, Adobe Lightroom Classic, and Photos for macOS. In 2019, Luminar won a gold award at the Digital Camera Grand Prix for its innovation in the photo editing space. Luminar has also been recognized as the Best Photo Software Product in the 2019 and 2020 by EISA awards. The application has filters, which can be used to edit images. The effects can be combined using layers and masks. Luminar works directly with raw files. The latest version to be currently available is Luminar AI, which was released September 2020. History Macphun launched Luminar in November 2016. An update with new tools and features called Luminar Neptune was released in 2017. Initially developed only for macOS, Macphun launched Luminar 2018 for Windows PC version in late 2017. At the same time, the company announced that it will change its name to Skylum. Luminar 2018, which was launched in November 2017, introduced a RAW develop module, changed the user interface and also added new filters. Reception Luminar has generally received positive reviews. Tech Radar reviewed the beta version of the application in 2016 and gave it a 4.5 out of 5. MyMac Reviews rated it 9 out of 10. Digital Rev appreciated the number of features offered by the software writing that "amateur Instagrammers and pro-photogs alike will feel like they’re in an Aladdin’s cave of wonders with what’s on offer in that regard," but pointed out that "in general Luminar does suffer from some problems when greater editing functions are stacked up against Lightroom or Photoshop’s." In 2020, Macworld gave Luminar a rating of 4.5 out of 5 and named it an Editors’ Choice, praising its “breakthrough AI Sky Replacement feature” and saying that it offers “intuitive, easy-to-use image editing.” Digital Camera World gave Luminar 4 a rating of 4 out of 5 stars, pointing out as advantages its integrated image catalog, non-destructive editing, AI Sky Replacement tool, and included plug-in version. Ghacks gave it a four out of five. Writing about Luminar 2018, Digital Trends expressed that it "is an excellent photo editor for beginners and hobbyists, with enough features to appease even some professional photographers." Appreciating the simple interface of the software, TechRepublic wrote that "Skylum's Luminar 2018, on the other hand, addresses common image issues—including color correction, brightening and darkening, cropping, and applying filters—without requiring you to become an expert in manipulating a variety of palettes, layers, and tools." In 2017 Luminar and in 2019, Luminar Flex was named the Best Software Plugin in the Lucie Technical Awards. In the same years, Luminar was awarded the Best Imaging Software by TIPA. In 2018, Luminar was named an Editor’s Pick by Outdoor Photographer. In October 2019, nature photography website Nature TTL reviewed the Luminar 4 software and gave it 5 out of 5. References Photo software Raster graphics editors HDR tone mapping software 2017 software
5071308
https://en.wikipedia.org/wiki/Sudbury%20and%20Copper%20Cliff%20Suburban%20Electric%20Railway
Sudbury and Copper Cliff Suburban Electric Railway
The Sudbury and Copper Cliff Suburban Electric Railway was a privately-owned electric interurban railway and streetcar system that connected several neighbourhoods in the town (and later city) of Sudbury, along with what was then the town of Copper Cliff in Ontario, Canada. Incorporated in 1912, the company began revenue passenger service in November 1915. It ultimately grew into a system of three radial lines which joined together in downtown Sudbury. By the late 1940s, however, rail service was being supplemented with diesel buses. Rail service ended in October 1950 and was completely replaced with bus service, which over time evolved into municipal Sudbury Transit and later GOVA services. History The early history of the railway was coloured by concerns about Copper Cliff becoming a suburb of Sudbury. An earlier 1906 plan for a Sudbury, Copper Cliff, and Creighton Electric Railway that would have connected the two centres with the now-abandoned mining town of Creighton had been discarded after heavy opposition from Copper Cliff merchants, who feared the decline of the town's commercial core due to competition with Sudbury, which had over six times as many businesses. The same merchants also opposed the new plan, but were overruled by a vote by Copper Cliff ratepayers. Early on, the railway company was run by Noël Desmarais, a local Franco-Ontarian businessman and grandfather of the financier Paul Desmarais, who later owned and managed the company. It began operations in 1912 with a single car that was borrowed from the Toronto Suburban Railway. The car was never returned and operated on the SCCSER until the early 1930s when it was sold for scrap. Service was only provided to Copper Cliff initially, with the lines on Notre Dame and Elgin opening shortly afterwards. The railway received electric power through the Wahnapitae Power Company, which was privately owned until it was acquired by the Hydro-Electric Power Commission of Ontario in 1930. The railway received a major boost in ridership by the decision of the Canadian Copper Company, after the instating of the eight-hour day, to limit the growth of Copper Cliff in favour of a workforce that commuted from Sudbury. It was also around this time that the small constellation of informal industrial villages such as East Smelter, Orford Village, and Shantytown disappeared or were consolidated into the urban centre of Copper Cliff. In Sudbury, the population of certain ethnic groups, such as Finns, began to grow, as immigrant mine and smelter workers could now live in Sudbury and commute to work at the industrial operations around Copper Cliff. At the time, the Sudbury Star reported that while in 1915, only about twenty workers commuted from Sudbury to Copper Cliff, a year later that number had risen to more than two hundred. On the far west end of Sudbury, the area around the Copper Cliff Road (later known as Lorne Street), which was also the route of the railway and which had previously been the site of the Gatchell family farm and dairy, was gradually subdivided and became the site of Sudbury's "first real suburban development." Gatchell became a predominantly Italian neighbourhood, one which mirrored Copper Cliff's Little Italy. Similarly, the West End neighbourhood grew extensively around this time along the Copper Cliff Road and Regent Street corridors. The railway company was not initially very financially successful, and was unable to pay dividends to its shareholders. In 1920, shareholders proposed that the town of Sudbury buy them out, which failed in a 4–2 town council vote. By 1928, an audit of the town's finances recommended that the town form a reserve in case the railway's debts to the town, which dated back to 1916, were realized as a loss. In 1943, the workers at the railway unionized, but by then its time was nearly over as the SCCSER began buying buses in 1947. The line to Bell Park was the first to close in 1948 with the two other lines ceasing service in 1950. The SCCSER rolling stock sat at the carhouse for more than two years, finally being cut up for scrap in 1953. The SCCSER was reorganized in 1951 and renamed Sudbury Bus Lines Limited. Shortly after, it was purchased for a symbolic $1 by the financier Paul Desmarais, who used it as a platform for buying up a number of bus lines throughout Ontario, allowing him to launch his business empire. Bus operations would continue under the new name until 1966, when the company ceased operations and was replaced by Laurentian Transit, which was a consortium of local private bus companies. It in turn was supplanted by Sudbury Transit in 1972, the direct predecessor to the current GOVA municipal transit agency. Routes The SCCSER had three main routes radiating from Elm Street in downtown Sudbury. Western line The westerly (and busiest) route ran a short distance from the Elm/Durham intersection to Lorne Street, where it turned south and entered dedicated trackage along the east side of the road. After about , the streetcar line and the road angled southwest, running parallel to the Canadian Pacific Railway's Webbwood Subdivision. At Gatchell, the streetcar line split into two tracks opposite Tuddenham Street to facilitate the passing of cars in rush hour. The line, once again single-track, then continued southwest, passing underneath the INCO railway via a specially-constructed underpass to a junction near Balsam Street in Copper Cliff, where it met the shuttle to the Inco refinery a short distance away. The line then angled northwest and ended at the two-track station near downtown Copper Cliff. It operated right up until the end of rail service in 1950. In addition to this service there was also a short-turn service that ended at Gatchell during peak hours. Northern line The northerly route ran from a wye at the Elm/Durham intersection to the SCCSER carhouse at St. Charles Street. This line was entirely single-track and operated for a short distance in the middle of Notre Dame Avenue before entering its own right-of-way on the east side of the road parallel to the CP Stobie Spur. This line operated right up until the end of service in 1950, but had been replaced by buses during peak periods in its last months of operation to clear the line for cars deadheading to and from the carhouse. Eastern line The easterly line ran east in the middle of Elm Street for approximately before turning south on Lisgar Street, west on Cedar Street, and south again on Durham Street. This odd arrangement was rumoured to have existed in order to ensure the carline's passage by a liquor store owned by a controlling interest in the SCCSER, but was more likely done to avoid crossing the CP Stobie Spur twice. Coming to the end of the trackage on Durham, the line entered a private right of way on the south side of Elgin Street and angled southeast parallel to the CP mainline through downtown. Passing the downtown Sudbury CP station at Van Horne, the line turned south across the Nelson Street "Iron" bridge (which was repurposed into a pedestrian bridge and is still used as such today ) and continued in the middle of Nelson Street to John Street, turning east and running one block to Elizabeth Street, where it turned south once again and ran down the middle of Elizabeth to Bell Park on the shore of Ramsey Lake. This line was also entirely single-track and ceased service in 1948 after the tracks on Lisgar Street were accidentally paved over by a contractor doing road repairs there. The SCCSER today Today the SCCSER has all but vanished. None of the streetcar equipment was ever retained for preservation. The streetcar alignments on Lorne and Notre Dame have since been obliterated by road-widening projects and all track was removed from city streets. Traces of the right-of-way near Copper Cliff, including the INCO railway underpass are purported to exist, however they could not be found on any satellite photographs of the area. A fare box used on the railway has survived and is preserved as an artifact at the Flour Mill Museum in the same neighbourhood where the original car barns were located. See also List of street railways in Canada List of Ontario railways List of defunct Canadian railways Interurban Streetcar History of rail transport in Canada Public transport in Canada References Citations Bibliography Further reading External links Dave's Electric Railroads davesrailpix.com’s server IP address could not be found. Sudbury & Copper Cliff Suburban Electric Railway streetcar 1 davesrailpix.com’s server IP address could not be found. Sudbury & Copper Cliff Suburban Electric Railway streetcar 2 davesrailpix.com’s server IP address could not be found. Sudbury & Copper Cliff Suburban Electric Railway streetcar #34 at Copper Cliff davesrailpix.com’s server IP address could not be found. at 1900 15 Nov, 2018 Defunct Ontario railways Transit agencies in Ontario Rail transport in Greater Sudbury Passenger rail transport in Greater Sudbury Electric railways in Canada Interurban railways in Ontario Street railways in Ontario
35062272
https://en.wikipedia.org/wiki/Coupa
Coupa
Coupa Software is a global technology platform for Business Spend Management (BSM). The company is headquartered in San Mateo, California with offices throughout Europe, Latin America, and Asia Pacific. Coupa helps companies gain visibility into, and control over, the money and resources spent within their organizations. History Dave Stephens and Noah Eisner founded Coupa in 2006. Rob Bernshteyn took over as Coupa's CEO in February 2009. In 2014, The Forrester Wave named Coupa a market leader. In May, Ariba, a unit of SAP and a competitor of Coupa, filed a lawsuit alleging that Coupa misappropriated Ariba trade secrets. Also in 2014, the company expanded globally, announcing new operations in Dublin, Ireland. In February 2015, Coupa acquired the assets of ZenPurchase, an enterprise procurement software company, as well as acquiring InvoiceSmash, an e-invoicing vendor, and TripScanner, an open booking vendor, in July. Coupa was also positioned in the Leaders quadrant of the Gartner Magic Quadrant for Procure-to-Pay Suites for Indirect Procurement. On September 24, 2015, Ariba and Coupa settled their lawsuit. Coupa acknowledged its possession of Ariba information and is required to abide by certain procedures to prevent any misappropriation in the future. Coupa paid Ariba an undisclosed amount. In January 2016, Coupa acquired Contractually, a cloud-based contract management solution. In October, Coupa Software went public on the Nasdaq, trading as COUP. In 2020, Coupa acquired AI supply chain design company LLamasoft. On June 15th, 2021, Coupa launched its Sustainable Business Spend Management (BSM) platform, to help business owners with their sustainability initiatives. Technology Coupa's cloud-based Business Spend Management solution manages transactions across procurement, payment, and supply chain. Coupa is cloud-based. In 2006, its first free open-source product, Coupa Express, was built using RoR. In 2007, the company launched a SaaS (Software-as-a-service) product called Coupa On Demand for SMBs (small and midsize businesses). In November 2011, Coupa released its first spend analysis product called Coupa Spend Optimizer. In April 2014, Coupa introduced Coupa Inventory, a portal for viewing inventory availability when ordering, to reduce wasteful spending. With Coupa’s acquisition of LLamasoft in 2020, the company added AI-powered supply chain design to its platform. In 2021, Coupa added configurations to its BSM platform, embedding ESG to optimize supply chains and reduce CO2 emissions. Coupa's BSM platform collects anonymized customer data to make recommendations to improve business transactions. Customers and partners As of 2018, Coupa indicated it had 700+ customers worldwide in 40 countries and in 150 currencies, including Nike, Caterpillar, Slack, Coca-Cola Consolidated, Adidas, Amazon, BNP Paribas, TD Bank Salesforce.com, and Rent-A-Center. In 2011, Coupa started working with the Independent Purchasing Cooperative (IPC) of Subway, to help its franchises in 98 countries with its cost management. Salesforce.com has its sales managers use Coupa to monitor travel expenses and sales outcomes. In October 2014, Coupa and NetSuite (NYSE: N) announced a strategic partnership to jointly market each other's products. References External links How Coupa’s savings-as-a-service model is paying off all round, Hot Topics, 2015 2006 establishments in California American companies established in 2006 Software companies established in 2006 Software companies based in the San Francisco Bay Area Companies based in San Mateo, California Companies listed on the Nasdaq 2016 initial public offerings Software companies of the United States CRM software companies
11563117
https://en.wikipedia.org/wiki/Distinguishing%20attack
Distinguishing attack
In cryptography, a distinguishing attack is any form of cryptanalysis on data encrypted by a cipher that allows an attacker to distinguish the encrypted data from random data. Modern symmetric-key ciphers are specifically designed to be immune to such an attack. In other words, modern encryption schemes are pseudorandom permutations and are designed to have ciphertext indistinguishability. If an algorithm is found that can distinguish the output from random faster than a brute force search, then that is considered a break of the cipher. A similar concept is the known-key distinguishing attack, whereby an attacker knows the key and can find a structural property in cipher, where the transformation from plaintext to ciphertext is not random. Overview To prove that a cryptographic function is safe, it is often compared to a random oracle. If a function would be a random oracle, then an attacker is not able to predict any of the output of the function. If a function is distinguishable from a random oracle, it has non-random properties. That is, there exists a relation between different outputs, or between input and output, which can be used by an attacker for example to find (a part of) the input. Example Let T be a sequence of random bits, generated by a random oracle and S be a sequence generated by a pseudo-random bit generator. Two parties use one encryption system to encrypt a message M of length n as the bitwise XOR of M and the next n bits of T or S respectively. The output of the encryption using T is truly random. Now if the sequence S cannot be distinguished from T, the output of the encryption with S will appear random as well. If the sequence S is distinguishable, then the encryption of M with S may reveal information of M. Two systems S and T are said to be indistinguishable if there exists no algorithm D, connected to either S or T, able to decide whether it is connected to S or T. A distinguishing attack is given by such an algorithm D. It is broadly an attack in which the attacker is given a black box containing either an instance of the system under attack with an unknown key, or a random object in the domain that the system aims to emulate, then if the algorithm is able to tell whether the system or the random object is in the black box, one has an attack. For example, a distinguishing attack on a stream cipher such as RC4 might be one that determines whether a given stream of bytes is random or generated by RC4 with an unknown key. Examples Classic examples of distinguishing attack on a popular stream cipher was by Itsik Mantin and Adi Shamir who showed that the 2nd output byte of RC4 was heavily biased toward zero. In another example, Souradyuti Paul and Bart Preneel of COSIC have shown that the XOR value of the 1st and 2nd outputs of RC4 is also non-uniform. Significantly, both the above theoretical biases can be demonstrable through computer simulation. See also Randomness test References External links Source Indifferentiability Cryptographic attacks
318553
https://en.wikipedia.org/wiki/Ten%C3%A9s%20Empanadas%20Graciela
Tenés Empanadas Graciela
Tenés Empanadas Graciela (short TEG) is a turn-based strategy game distributed by several popular Linux distributions. The idea for this free and open-source software program came from the board game TEG, which itself is based on the popular Risk game but differs in many aspects of the rules. Gameplay Using a Server several players can fight each other for world domination and of course also chat together. Various maps are used as battlefield and also several languages like Spanish, German, French and Polish are available. History In 1996 the Argentine Ricardo "riq" Quesada started developing the game, and released it under the GNU GPL-2-0-only. He stopped working on it for a few years but brought the project back to life in early 2000. Some developers, graphic artists and translators joined the project located at SourceForge's servers and made it a success. TEG was an early supporter of the FOSS multiplayer client-server system GGZ Gaming Zone. Around 2014 the project moved from Sourceforge to GitHub. Reception "Tenés empanadas Graciela" was reviewed as a notable Risk clone several times by free and open-source software associated media outlets, such as in 2002 by the Linux Journal and Linux Magazine. In 2007 Linux.com reviewed "Tenés empanadas Graciela" again as notable Risk clone. The Linux Game Tome lists the game with 4.2 of 5 stars. The game was reviewed in 2012 by republica.com. Versions and ports The game is widely distributed by several Linux distributions like Ubuntu, Debian, Gentoo Linux, Arch Linux and ported to other Operating Systems like macOS. There is also an Argentine online version of the game called WebTeg which allows users to play with just a web browser. See also Lux, a series of proprietary Risk-clones that work on Linux. List of open source games References External links TEG Old official page on SourceForge.net teg Official page on GitHub Web TEG Gagné, Marcel, (2002), "Networking for Pleasure" Linux Journal Greve, Georg CF, (2002), "Brave GNU World", Linux Magazine, Issue 22. (archived) Linux games Open-source video games Strategy video games Freeware games MacOS games
50879654
https://en.wikipedia.org/wiki/Sunway%20TaihuLight
Sunway TaihuLight
The Sunway TaihuLight (, Shénwēi·tàihú zhī guāng) is a Chinese supercomputer which, , is ranked fourth in the TOP500 list, with a LINPACK benchmark rating of 93 petaflops. The name is translated as divine power, the light of Taihu Lake. This is nearly three times as fast as the previous Tianhe-2, which ran at 34 petaflops. , it is ranked as the 16th most energy-efficient supercomputer in the Green500, with an efficiency of 6.051 GFlops/watt. It was designed by the National Research Center of Parallel Computer Engineering & Technology (NRCPC) and is located at the National Supercomputing Center in Wuxi in the city of Wuxi, in Jiangsu province, China. The Sunway TaihuLight was the world's fastest supercomputer for two years, from June 2016 to June 2018, according to the TOP500 lists. The record was surpassed in June 2018 by IBM's Summit. Architecture The Sunway TaihuLight uses a total of 40,960 Chinese-designed SW26010 manycore 64-bit RISC processors based on the Sunway architecture. Each processor chip contains 256 processing cores, and an additional four auxiliary cores for system management (also RISC cores, just more fully featured) for a total of 10,649,600 CPU cores across the entire system. The processing cores feature 64 KB of scratchpad memory for data (and 16 KB for instructions) and communicate via a network on a chip, instead of having a traditional cache hierarchy. Software The system runs on its own operating system, Sunway RaiseOS 2.0.5, which is based on Linux. The system has its own customized implementation of OpenACC 2.0 to aid the parallelization of code. Future development China's first exascale supercomputer was scheduled to enter service by 2020 according to the head of the school of computing at the National University of Defense Technology (NUDT). According to the national plan for the next generation of high performance computers, the country would have develop an exascale computer during the 13th Five-Year-Plan period (2016–2020). The government of Tianjin Binhai New Area, NUDT and the National Supercomputing Center of Tianjin are working on the project. The investment is likely to hit 3 billion yuan ($470.6 million). See also Sunway BlueLight Manycore processor Massively parallel processor array Supercomputing in China Summit (supercomputer) References External links Top500 list entry for the Sunway TaihuLight CCTV video news story on Sunway TaihuLight Hardware of Sunway TaihuLight - BBC 5-minute video 2016 in technology Petascale computers Supercomputers Supercomputing in China 64-bit computers
21729503
https://en.wikipedia.org/wiki/IPL%20Information%20Processing%20Limited
IPL Information Processing Limited
IPL Information Processing Limited, commonly known as IPL, is a privately owned European software services company headquartered in Bath, UK, providing business consultancy, technical consultancy, IT solutions and support services. The firm was founded in 1979 and employs 278 staff. For the year ended 30 September 2014, the company posted a turnover of £27.3m. History IPL was founded in 1979 as a software development company by Michael Johnson, David Embleton and Clive Jones. Over time, it has expanded into the consultancy and managed services markets. It moved to its current headquarters on Grove Street in 1986, which it extended in 1992. The company underwent a management buyout in 2008, which saw Shaun Davey become CEO. In March 2012, IPL sold its Testing Products division, which had developed and distributed the Cantata++ and AdaTEST software testing tools, to QA Systems. In October 2012, Neil Ellett replaced Shaun Davey as CEO. Ellett was replaced in June 2013 by Paul Jobbins. In April 2016, IPL was acquired by Civica and as of the January 2017, has been rebranded as Civica Digital. Operations The company operates three main departments: Consultancy (branded 'Look'), Delivery (branded 'Build') and Service (branded 'Improve'). It also has an Innovations & Solutions Unit, which looks after research and development and the company's propositions and intellectual property. The company operates in the following sectors: Transport Financial services Oil & Gas Pharmaceuticals & Life sciences Central government Local government Emergency services Defence Notable clients include FirstGroup, Rail Settlement Plan, Highways England, the Environment Agency, Kent Police, the Home Office and Nationwide Building Society. Certifications IPL's Quality Management System has been certified to ISO 9001 since 1986, and the TickIT scheme since 1992. The company upgraded its TickIT certification to the new TickITplus standard in 2012. IPL holds ISO/IEC 27001:2013 (Information Security) and ISO 14001 (Environmental) certifications. The company is also a recognised Investor in People. Community involvement IPL was the main sponsor of Bath Rugby during the 2009–10 and 2010–11 seasons. It is now the club's Official Business Partner, and the South Stand at the Recreation Ground is named the IPL Stand. The company has also provided support for the Bath Rugby Foundation. IPL also sponsored the University of Bath's TeamBath rugby team from 2011 to 2013, and the netball team from 2011 to 2013. The company sponsored the 2013 Bath Digital Festival, and also sponsors an undergraduate academic award at the University of Bristol. References External links IT Network Support Software companies of the United Kingdom Management consulting firms of the United Kingdom Technology companies of the United Kingdom Information technology consulting firms Information technology companies of the United Kingdom Mass media technology Defence companies of the United Kingdom Government software Companies based in Bath, Somerset Transport software
14260687
https://en.wikipedia.org/wiki/Mobile%20operating%20system
Mobile operating system
A mobile operating system is an operating system for mobile phones, tablets, smartwatches, 2-in-1 PCs, smart speakers, or other mobile devices. While computers such as typical laptops are 'mobile', the operating systems used on them are generally not considered mobile ones, as they were originally designed for desktop computers that historically did not have or need specific mobile features. This distinction is becoming blurred in some newer operating systems that are hybrids made for both uses. Mobile operating systems combine features of a personal computer operating system with other features useful for mobile or handheld use, and usually including a wireless inbuilt modem and SIM tray for telephony and data connection. By Q1 2018, over 383 million smartphones were sold with 86.2 percent running Android and 12.9 percent running iOS. Android alone is more popular than the popular desktop operating system Microsoft Windows, and in general smartphone use (even without tablets) outnumbers desktop use. Mobile devices, with mobile communications abilities (e.g., smartphones), contain two mobile operating systemsthe main user-facing software platform is supplemented by a second low-level proprietary real-time operating system which operates the radio and other hardware. Research has shown that these low-level systems may contain a range of security vulnerabilities permitting malicious base stations to gain high levels of control over the mobile device. Mobile operating systems have majority use since 2017 (measured by web use); with even only the smartphones running them (excluding tablets) having majority use, more used than any other kind of device. Thus traditional desktop OS is now a minority-used kind of OS; see usage share of operating systems. However, variations occur in popularity by regions, while desktop-minority also applies on some days in countries such as United States and United Kingdom. Timeline Mobile operating system milestones mirror the development of mobile phones, PDAs, and smartphones: Pre-1993 1973–1993 – Mobile phones use embedded systems to control operation. 1993–1999 1993 April – PenPoint OS by GO Corporation becomes available on the AT&T EO Personal Communicator. August – Apple launches Newton OS running on their Newton series of portable computers. 1994 March – Magic Cap OS by General Magic is first introduced on the Sony Magic Link PDA. August – The first smartphone, the IBM Simon, has a touchscreen, email, and PDA features. 1996 March – The Palm Pilot 1000 personal digital assistant is introduced with the Palm OS mobile operating system. August – Nokia releases the Nokia 9000 Communicator running an integrated system based on the PEN/GEOS 3.0 OS from Geoworks. 1997 – EPOC32 first appears on the Psion Series 5 PDA. Release 6 of EPOC32 will later be renamed to Symbian OS. 1998 – Symbian Ltd. is formed as a joint venture by Psion, Ericsson, Motorola, and Nokia, Psion's EPOC32 OS becomes Symbian's EPOC operating system, and is later renamed to Symbian OS. Symbian's OS was used by those companies and several other major mobile phone brands, but especially Nokia. 1999 June – Qualcomm's pdQ becomes the first smartphone with Palm OS. October – Nokia S40 Platform is officially introduced along with the Nokia 7110, the first phone with T9 Predictive text input and a Wireless Application Protocol (WAP) browser for accessing specially formatted Internet data. 2000s 2000 – The Ericsson R380 is released with EPOC32 Release 5, marking the first use on a phone of what's to become known as Symbian OS (as of Release 6). 2001 June – Nokia's Symbian Series 80 platform is first released on the Nokia 9210 Communicator This is the first phone running an OS branded as Symbian, and the first phone using that OS that allows user installation of additional software. September – Qualcomm's Binary Runtime Environment for Wireless (BREW) platform on their REX real-time operating system (RTOS) is first released on the Kyocera QCP-3035. 2002 March BlackBerry releases its first smartphone, running Java 2 Micro Edition (J2ME). UIQ is first released, at v2.0, on Symbian OS, and becomes available later in the year on the Sony Ericsson P800, the successor to the Ericsson R380. June Microsoft's first Windows CE (Pocket PC) smartphones are introduced. Nokia's Symbian Series 60 (S60) platform is released with the Nokia 7650, Nokia's first phone with a camera and Multimedia Messaging Service (MMS). S60 would form the basis of the OS on most of Nokia's smartphones until 2011, when they adopted Microsoft's Windows Phone 7. S60 was also used on some phones from Samsung and others, and later by Sony Ericsson after the consolidation of different Symbian UI variants in 2008. October – The Danger Hiptop (T-Mobile Sidekick in U.S.) is first released by Danger, Inc., running DangerOS. 2003 – Motorola introduces first linux based cellphone Motorola A760 base on Linux MontaVista distribution. 2005 May – Microsoft announces Windows Mobile 5.0. November – Nokia introduces Maemo OS on the first, small Internet tablet, the N770, with a 4.13" screen. 2007 January – Apple's iPhone with iOS (named "iPhone OS" for its first three releases) is introduced as a "widescreen iPod," "mobile phone," and "Internet communicator". February – Microsoft announces Windows Mobile 6.0. May – Palm announces the Palm Foleo, a "Mobile Companion" device similar to a subnotebook computer, running a modified Linux kernel and relying on a companion Palm Treo smartphone to send and retrieve mail, as well as provide data connectivity when away from WiFi. Palm canceled Foleo development on September 4, 2007, after facing public criticism. June - World's very first iPhone is released in the United States. November – Open Handset Alliance (OHA) is established, led by Google with 34 members (HTC, Sony, Dell, Intel, Motorola, Samsung, LG, etc.) 2008 February – LiMo Foundation announces the first phones running the LiMo mobile Linux distribution, from Motorola, NEC, Panasonic Mobile, and Samsung, released later in the year. The LiMo Foundation later became the Tizen Association and LiMo was subsumed by Tizen. June – Nokia becomes the sole owner of Symbian Ltd. The Symbian Foundation was then formed to co-ordinate the future development of the Symbian platform among the corporations using it, in a manner similar to the Open Handset Alliance with Android. Nokia remained the major contributor to Symbian's code. July – Apple releases iPhone OS 2 with the iPhone 3G, making available Apple's App Store. October – OHA releases Android (based on Linux kernel) 1.0 with the HTC Dream (T-Mobile G1) as the first Android phone. November – Symbian^1, the Symbian Foundation's touch-specific S60-based platform (equivalent to S60 5th edition) is first released on Nokia's first touchscreen Symbian phone, the Nokia 5800 XpressMusic, with a resistive screen and a stylus. Symbian^1 being derived from S60 meant that support for UIQ disappeared and no further devices using UIQ were released. 2009 January Intel announces Moblin 2, specifically created for netbooks that run the company's Atom processor. In April 2009 Intel turned Moblin over to the Linux Foundation. Palm introduces webOS with the Palm Pre (released in June). The new OS is not backwards compatible with their previous Palm OS. February Palm announces that no further devices with Palm OS are going to be released by the company. (The last was the Palm Centro, released October 14, 2007.) Microsoft announces Windows Mobile 6.5, an "unwanted stopgap" update to Windows Mobile 6.1 intended to bridge the gap between version 6.1 and the then yet-to-be released Windows Mobile 7 (later canceled in favor of Windows Phone 7). The first devices running it appeared in late October 2009. May – DangerOS 5.0 becomes available, based on NetBSD. June – Apple releases iPhone OS 3 with the iPhone 3GS. November – Nokia releases the Nokia N900, its first and only smartphone running the Maemo OS intended for "handheld computers...with voice capability," while stating that they remain focused on Symbian S60 as their smartphone OS. (Nokia had previously released three Mobile Internet devices running Maemo, without cellular network connectivity.) 2010s 2010 February MeeGo is announced, a mobile Linux distribution merging Maemo from Nokia and Moblin from Intel and Linux Foundation, to be hosted by Linux Foundation. MeeGo is not backwards compatible with any previous operating system. Samsung introduces the Bada OS and shows the first Bada smartphone, the Samsung S8500. It's later released in May 2010. April Apple releases the iPad (first generation) with iPhone OS 3.2. This is the first version of the OS to support tablet computers. For its next major version (4.0) iPhone OS will be renamed iOS. HP acquires Palm in order to use webOS in multiple new products, including smartphones, tablets, and printers, later stating their intent to use it as the universal platform for all their devices. May – Microsoft Kin phone line with KIN OS (based on Windows CE and a "close cousin" to Windows Phone) become available. June – Apple releases iOS 4, renamed from iPhone OS, with the iPhone 4. July – Microsoft Kin phones and KIN OS are discontinued. September Apple releases a variant of iOS powering the new 2nd generation Apple TV. Symbian^3 is first released on the Nokia N8. This would be Nokia's last flagship device running Symbian (though not their last Symbian phone), before switching to Windows Phone 7 for future flagship phones. The Danger Hiptop line and DangerOS are discontinued as a result of Microsoft's acquisition of Danger, Inc. in 2008. November Nokia assumes full control over Symbian as the Symbian Foundation disintegrates. Windows Phone OS is released on Windows Phone 7 phones by HTC, LG, Samsung, and Dell. The new OS is not backwards compatible with the prior Windows Mobile OS. 2011 February Android 3.0 (Honeycomb), the first version to officially support tablet computers, is released on the Motorola Xoom. Nokia abandons the Symbian OS and announces that it would use Microsoft's Windows Phone 7 as its primary smartphone platform, while Symbian would be gradually wound down. April – BlackBerry Tablet OS, based on QNX Neutrino is released on the BlackBerry PlayBook. July Mozilla announces their Boot to Gecko project (later named Firefox OS) to develop an OS for handheld devices emphasizing standards-based Web technologies, similar to webOS. webOS 3.0, the first version to support tablet computers, is released on the HP TouchPad. August – HP announces that webOS device development and production lines would be halted. The last HP webOS version, 3.0.5, is released on January 12, 2012. September MeeGo is introduced with the limited-release Nokia N9, Nokia's first and only consumer device to use the OS. (A small number of the Nokia N950, a MeeGo phone available only to developers, were released in mid-2011.) After Nokia's abandonment of MeeGo, Intel and the Linux Foundation announce a partnership with Samsung to launch Tizen, shifting their focus from MeeGo (Intel and Linux Foundation) and Bada (Samsung) during 2011 and 2012. October Apple releases iOS 5 with the iPhone 4S, integrating the Siri voice assistant. The Mer project is announced, based on an ultra-portable core for building products, composed of Linux, HTML5, QML, and JavaScript, which is derived from the MeeGo codebase. November – Fire OS, a fork of the Android operating system, is released by Amazon.com on the Kindle Fire tablet. 2012 May – Nokia releases the Nokia 808 PureView, later confirmed (in January 2013) to be the last Symbian smartphone. This phone was followed by a single last Symbian software update, "Nokia Belle, Feature Pack 2," later in 2012. July Finnish start-up Jolla, formed by former Nokia employees, announces that MeeGo's community-driven successor Mer would be the basis of their new Sailfish smartphone OS. Mozilla announces that the project formerly named Boot to Gecko (which is built atop an Android Linux kernel using Android drivers and services; however it uses no Java-like code of Android) is now Firefox OS (since discontinued) and has several handset OEMs on board. August – Samsung announces they won't ship further phones using their Bada OS, instead focusing on Windows Phone 8 and Android. September – Apple releases iOS 6 with the iPhone 5. 2013 January – BlackBerry releases their new operating system for smartphones, BlackBerry 10, with their Q10 and Z10 smartphones. BlackBerry 10 is not backwards compatible with the BlackBerry OS used on their previous smartphones. February – HP sells webOS to LG. September – Apple releases iOS 7 with the iPhone 5S and iPhone 5C. October Canonical announces Ubuntu Touch, a version of the Linux distribution expressly designed for smartphones. The OS is built on the Android Linux kernel, using Android drivers and services, but does not use any of the Java-like code of Android. Google releases Android KitKat 4.4. November – Jolla releases Sailfish OS on the Jolla smartphone. 2014 February Microsoft releases Windows Phone 8.1 Nokia introduces their Nokia X platform OS as an Android 4.1.2 Jelly Bean fork on the Nokia X family of smartphones. Similar to Amazon.com's Fire OS, it replaces Google's apps and services with ones from Nokia (such as HERE Maps, Nokia Xpress and MixRadio, and Nokia's own app store) and Microsoft (such as Skype and Outlook), with a user interface that mimics the Windows Phone UI. After the acquisition of Nokia's devices unit, Microsoft announced in July 2014 that no more Nokia X smartphones would be introduced, marking the end of the platform just a few months later. August – The Samsung SM-Z9005 Z is the first phone released running Tizen, with v2.2.1 of the OS. September Apple releases iOS 8 with the iPhone 6 and 6 Plus. BlackBerry releases BlackBerry 10 version 10.3 with integration with the Amazon Appstore November – Google releases Android 5.0 "Lollipop" 2015 February – Google releases Android 5.1 "Lollipop". April LG releases the LG Watch Urbane LTE smartwatch running "LG Wearable Platform OS" based on webOS. This is a version of their Android Wear OS-based LG Watch Urbane, with added LTE connectivity. watchOS, based on iOS, is released by Apple with the Apple Watch. September Apple releases iOS 9 with the iPhone 6S and 6S Plus, iPad Pro, and iPad Mini 4, plus watchOS 2. tvOS 9 is also made distinct from iOS, with its own App Store, launching with Apple TV 4th generation. Google releases Android 6.0 "Marshmallow". October – BlackBerry announces that there are no plans to release new APIs and software development kits for BlackBerry 10, and future updates would focus on security and privacy enhancements only. November – Microsoft releases Windows 10 Mobile. 2016 February – Microsoft releases the Lumia 650, their last Windows 10 Mobile phone before discontinuing all mobile hardware production the following year. July – The BlackBerry Classic, the last device to date running a BlackBerry OS is discontinued. While BlackBerry Limited claimed to still be committed to the BlackBerry 10 operating system, they have since only shipped Android devices after releasing the BlackBerry Priv, their first Android smartphone in November 2015. August Google posts the Fuchsia source code on GitHub. Google releases Android 7.0 "Nougat". September – Apple releases iOS 10 with the iPhone 7 and 7 Plus, and watchOS 3 with the Apple Watch Series 1 and 2. November Tizen releases Tizen 3.0. BlackBerry releases BlackBerry 10 version 10.3.3. 2017 April Development of Ubuntu Touch is transferred from Canonical Ltd. to the UBports Foundation Samsung officially launches Android-based Samsung Experience custom firmware starting with version 8.1 on Samsung Galaxy S8. May Samsung announces Tizen 4.0 at Tizen Developer Conference 2017. August Google releases Android 8.0 "Oreo". September Apple releases iOS 11 with the iPhone 8 and 8 Plus and iPhone X, and watchOS 4 with the Apple Watch Series 3. October Microsoft announces that Windows 10 Mobile development is going into maintenance mode only, ending the release of any new features or functionality due to lack of market penetration and resultant lack of interest from app developers, and releases the final major update to it, the "Fall Creators Update." Cherry Mobile release CherryOS based on Android 2018 February Samsung releases Samsung Experience 9.0 based on Android "Oreo" 8.0 globally to Samsung Galaxy S8 and S8+. March Google and partners officially launches Android Go (based on Android "Oreo" 8.1 but tailored for low-end devices) with Nokia 1, Alcatel 1X, ZTE Tempo Go, General Mobile 8 Go, Micromax Bharat Go and Lava Z50. Google releases Android "9" as a developer preview. April Microsoft release Windows 10 Version 1803 "April 2018 Update". May Huawei release LiteOS version 2.1. August Google releases Android 9.0 "Pie". UBPorts released Ubuntu Touch OTA-14, upgrading the OS based on the Canonical's long-term support version of Ubuntu 16.04 LTS "Xenial Xerus". Xiaomi officially introduces MIUI for POCO for their Poco series smartphone. Samsung officially introduces Tizen 4.0 with the release of Samsung Galaxy Watch series. September Apple releases iOS 12 with the iPhone XS and XS Max, and watchOS 5 with Apple Watch Series 4. Huawei releases EMUI 9.0. October Microsoft releases Windows 10 Version 1809 "October 2018 Update". November Samsung announces the One UI as the latest version of the Samsung Experience UI. Amazon released Fire OS 6 to supported Fire HD devices. 2019 January Microsoft announces that support for Windows 10 Mobile would end on December 10, 2019, and that Windows 10 Mobile users should migrate to iOS or Android phones. June Apple announces iOS 13, watchOS 6, and iPadOS as a distinct variant of iOS. August Huawei announced the EMUI 10 under beta release. September Apple releases iOS 13 with the iPhone 11 series, watchOS 6 with Apple Watch Series 5, and iPadOS with the 7th generation iPad. Google releases Android 10. The Librem 5, the first phone running PureOS, is released. October Samsung announces the One UI 2.0 as the latest version of their Galaxy Smartphone and Smartwatch UI. November Microsoft release the Windows 10 November 10, 2019 Update. Current software platforms These operating systems often run atop baseband or other real time operating systems that handle hardware aspects of the phone. Android Android (based on the modified Linux kernel) is a mobile operating system developed by Google. The base system is open-source (and only the kernel copyleft), but the apps and drivers which provide functionality are increasingly becoming closed-source. Besides having the largest installed base worldwide on smartphones, it is also the most popular operating system for general purpose computers (a category that includes desktop computers and mobile devices), even though Android is not a popular operating system for regular (desktop) personal computers (PCs). Although the Android operating system is free and open-source software, in devices sold, much of the software bundled with it (including Google apps and vendor-installed software) is proprietary software and closed source. Android's releases before 2.0 (1.0, 1.5, 1.6) were used exclusively on mobile phones. Android 2.x releases were mostly used for mobile phones but also some tablets, Android 3.0 was a tablet-oriented release and does not officially run on mobile phones, while both phone and tablet compatibility was merged with Android 4.0. The current Android version is Android 12, released on October 4, 2021. ==== Android One ==== Android One is a software experience that runs on the unmodified Android operating system, which closely resembles those running on Pixel devices or previously, the Google Nexus program. Unlike most of the "stock" Androids running on the market, Android One UI closely resembles the Pixel UI, due to Android One being a software experience developed by Google and distributed to partners who signup for the program, such as Nokia Mobile(HMD) and Xiaomi. Thus, the overall UI is intended to be as clean as possible. OEM partners may tweak or add additional apps such as cameras to the firmware, otherwise most of the apps will be handled by Google proprietary apps. The update was handled by Google and will be internally tested by OEM before being distributed via OTA update to the end users. Android One version are same as AOSP, starting from Android 5.0 "Lollipop" BlackBerry Secure BlackBerry Secure is an operating system developed by BlackBerry, based on the Android Open Source Project (AOSP). Officially announced the name for their Android based front-end touch interface in August 2017, before the announcement, BlackBerry Secure was running on BlackBerry brand devices such as BlackBerry Priv, DTEK 50/60 and BlackBerry KeyOne. Currently, BlackBerry plan to license out the BlackBerry Secure to other OEM's. Current BlackBerry Secure version list BlackBerry Secure version 1.x (based on Android "Marshmallow" 6.x and "Nougat" 7.x) (initial release and minor UI update for devices that support Android Nougat) ColorOS ColorOS is a custom front-end touch interface, based on the Android Open Source Project (AOSP) and is developed by OPPO Electronics Corp. , OPPO officially releases ColorOS with every OPPO and Realme devices(Future Realme devices will be having their own custom version of ColorOS), and released an official ROM for the OnePlus One. CopperheadOS CopperheadOS, a security hardened version of Android EMUI Huawei EMUI is a front-end touch interface developed by Huawei Technologies Co. Ltd. and its sub-brand Honor which is based on Google's Android Open Source Project (AOSP). EMUI is preinstalled on most Huawei and Honor devices. While it was based on open source Android operating system, it consists of closed source and proprietary software. /e/ /e/ is an operating system forked from the source code of LineageOS (based on Android). /e/ targets Android smart phone devices, and uses MicroG as a replacement for Google Play Services. Current /e/ OS version list /e/ OS beta v0.1 (based on Android "Nougat" 7.1.x) /e/ OS beta v0.2 (based on Android "Oreo" 8.x.x) Fire OS Amazon Fire OS is an Android-based mobile operating system produced by Amazon for its Fire range of tablets, Echo and Echo Dot, and other content delivery devices like Fire TV (Previously for their Fire Phone). It is forked from Android. Fire OS primarily centers on content consumption, with a customized user interface and heavy ties to content available from Amazon's own storefronts and services. Current Fire OS version list Fire OS 1.x Fire OS 2.x Fire OS 3.x Fire OS 4.x Fire OS 5.x Fire OS 6.x Fire OS 7.x Flyme OS Flyme OS is an operating system developed by Meizu Technology Co., Ltd., an open source OS based on Google Android Open Source Project (AOSP). Flyme OS is mainly installed on Meizu Smartphones such as the MX's series; however, it also has official ROM support for a few Android devices. Current Flyme OS version list Flyme OS 1.x.x (based on Android "Ice Cream Sandwich" 4.0.3, initial release) Flyme OS 2.x.x (based on Android "Jelly Bean" 4.1.x – 4.2.x) Flyme OS 3.x.x (based on Android "Jelly Bean" 4.3.x) Flyme OS 4.x.x (based on Android "KitKat" 4.4.x) Flyme OS 5.x.x (based on Android "Lollipop" 5.0.x – 5.1.x) Flyme OS 6.x.x (based on Android "Nougat" 7.x, "Marshmallow" 6.0.x and Android "Lollipop" 5.0.x – 5.1.x for old devices) Flyme OS 7.x.x (based on Android "Pie" 9.0, "Oreo" 8.x and Android "Nougat" 7.x) Flyme OS 8.x.x (based on Android 10, "Pie" 9.0, "Oreo" 8.x and Android "Nougat" 7.x) GrapheneOS GrapheneOS, formerly named Android Hardening, is a variant of Android that runs on Pixel 2 or Pixel 3 hardware, is mainly developed by Daniel Micay , and aims to focus on security and privacy. HTC Sense HTC Sense is a software suite developed by HTC, used primarily on the company's Android-based devices. Serving as a successor to HTC's TouchFLO 3D software for Windows Mobile, Sense modifies many aspects of the Android user experience, incorporating added features (such as an altered home screen and keyboard), widgets, HTC-developed applications, and redesigned applications. The first device with Sense, the HTC Hero, was released in 2009. HTC Sense 1.x (based on Android "Eclair" 2.0/2.1, initial release) HTC Sense 2.x (based on Android "Eclair", "Froyo" and "Gingerbread" 2.0/2.1, 2.2.x and 2.3.x, redesigned UI) HTC Sense 3.x (based on Android "Gingerbread" 2.3.x, redesigned UI) HTC Sense 4.x (based on Android "Ice Cream Sandwich" and "Jelly Bean" 4.0.x and 4.1.x, redesigned UI) HTC Sense 5.x (based on Android "Jelly Bean" 4.1.x – 4.3.x, redesigned UI) HTC Sense 6.x (based on Android "KitKat" 4.4.x, redesigned UI) HTC Sense 7.x (based on Android "Lollipop" 5.0.x, redesigned UI) HTC Sense 8.x (based on Android "Marshmallow" 6.0.x, redesigned UI) HTC Sense 9.x (based on Android "Nougat" 7.x, redesigned UI) HTC Sense 10.x (based on Android "Oreo" 8.x and "Pie" 9.0, redesigned UI) iQOO UI iQOO UI is a customer user interface that is based on Vivo Funtouch OS, which itself is based on the Android Open Source Project (AOSP). The overall UI mostly resemble its predecessor, however with their own customized UI on top of the Funtouch OS. Current iQOO UI version list iQOO UI 1.x - Based on Funtouch OS Indus OS Indus OS is a custom mobile operating system based on the Android Open Source Project (AOSP). It is developed by the Indus OS team based in India. No longer valid as of 2018Indus OS is available on Micromax, Intex, Karbonn, and other Indian smartphone brands. Current Indus OS version list Firstouch OS (based on Android "Lollipop" 5.0) Indus OS 2.0 (based on Android "Marshmallow" 6.0) Indus OS 3.0 (based on Android "Nougat" 7.0.1) LG UX LG UX (formerly named Optimus UI) is a front-end touch interface developed by LG Electronics with partners, featuring a full touch user interface. It is sometimes incorrectly identified as an operating system. LG UX is used internally by LG for sophisticated feature phones and tablet computers, and is not available for licensing by external parties. Optimus UI 2 which based on Android 4.1.2 has been released on the Optimus K II and the Optimus Neo 3. It features a more refined user interface compared to the prior version based on Android 4.1.1, would include together which new functionality such as voice shutter and quick memo. Current LG UX version list Optimus UI 1.x – based on Android "Gingerbread" 2.3.x, initial release Optimus UI 2.x – based on Android "Ice Cream Sandwich" and "Jelly Bean" 4.0.x and 4.1.x – 4.3.x, redesigned UI LG UX 3.x – based on Android "KitKat" and "Lollipop" 4.4.x and 5.0.x, redesigned UI LG UX 4.x – based on Android "Lollipop" and "Marshmallow" 5.1.x and 6.0.x, redesigned UI LG UX 5.x – based on Android "Marshmallow" and "Nougat" 6.0.x and 7.0.x, redesigned UI LG UX 6.x – based on Android "Nougat" 7.0.x, redesigned UI LG UX 6.x+ – based on Android "Oreo" 8.0.x, redesigned UI LG UX 7.x – based on Android "Oreo" 8.x, redesigned UI LG UX 7.x+ – based on Android "Oreo" 8.x., redesigned UI LG UX 8.x – based on Android "Pie" 9.0, redesigned UI LG UX 9.x – based on Android 10 redesigned UI LineageOS Lineage Android Distribution is a custom mobile operating system based on the Android Open Source Project (AOSP). It serves as the successor to the highly popular custom ROM, CyanogenMod, from which it was forked in December 2016 when Cyanogen Inc. announced it was discontinuing development and shut down the infrastructure behind the project. Since Cyanogen Inc. retained the rights to the Cyanogen name, the project rebranded its fork as LineageOS. Similar to CyanogenMod, it does not include any proprietary apps unless the user installs them. It allows Android users who can no longer obtain update support from their manufacturer to continue updating their OS version to the latest one based on official release from Google AOSP and heavy theme customization. MiFavor MiFavor, was the custom Android UI that was developed by ZTE for their smartphone that is running Android platform. Similar to most of the other Android UI, MiFavor replace most of the stock apps with ZTE owns apps, however the all UX still closely similar to the stock Android. Current MiFavor version list MiFavor 1.x – based on Android "KitKat" 4.4.x, initial release MiFavor 2.x – based on Android "Lollipop" 5.0.x – 5.1.x, redesigned UI MiFavor 3.x – based on Android "Marshmallow" 6.x, redesigned UI MiFavor 4.x – based on Android "Nougat" 7.x, redesigned UI MiFavor 5.x – based on Android "Oreo" 8.x, redesigned UI MiFavor 9.x – based on Android "Pie" 9.0, redesigned UI MiFavor 10.x – based on Android 10, redesigned UI MIUI Mi User Interface (MIUI), developed by the Chinese electronic company Xiaomi Inc., is a mobile operating system based on the Android Open Source Project (AOSP). MIUI is mostly found in Xiaomi smartphones such as the Mi and Redmi Series; however, it also has official ROM support for few Android devices. Although MIUI is based on AOSP, which is open source, it consists of closed source and proprietary software of its own. One UI One UI (formerly called TouchWiz and Samsung Experience) is a front-end touch interface developed by Samsung Electronics In 2008 with partners, featuring a full touch user interface. It is sometimes incorrectly identified as an independent operating system. Samsung Experience is used internally by Samsung for smartphones, feature phones and tablet computers, and is not available for licensing by external parties as it is closed source and proprietary. The Android version of Samsung Experience also comes with Samsung-made apps preloaded (except starting with the Galaxy S6 which have removed all Samsung pre-loaded apps installed, leaving one with Samsung Galaxy Store (formerly Galaxy Apps), to save storage space and initially due to the removal of MicroSD). With the release of Samsung Galaxy S8 and S8+, Samsung Experience 8.1 was preinstall on it with introducing new function known as Samsung DeX. Similar to the concept of Microsoft Continuum, Samsung DeX allowed high end Galaxy devices such as S8/S8+ or Note 8 to connect into a docking station, which extends the functionality to allow desktop-like functionality by connecting a keyboard, mouse, and monitor. Samsung also announced "Linux on Galaxy", which allows to use the standard Linux distribution on the DeX platform. Previous Samsung Android UI version list TouchWiz 3.x (based on Android 2.1 "Éclair" and Android 2.2 "Froyo") (Initial release for Android UI) TouchWiz 4.x (based on Android 2.3 "Gingerbread" and Android 3.0 "Honeycomb") (Minor UI update) TouchWiz Nature UX (based on Android 4.0 "Ice Cream Sandwich") (Minor UI update) TouchWiz Nature UX 2.x (based on Android 4.2 "Jellybean") (Minor UI update) TouchWiz Nature UX 3.x (based on Android 4.4 "KitKat") (Minor UI update) TouchWiz Nature UX 4.x (based on Android 5 "Lollipop") (Minor UI update) TouchWiz Nature UX 5.x (based on Android 5 "Lollipop") (Major UI update) TouchWiz Nature UX 6.x (based on Android 6 "Marshmallow") (Minor UI update) TouchWiz Grace UX (based on Android 6 "Marshmallow") (Major UI update) Samsung Experience 8.x (based on Android 7 "Nougat") (Initial release migrate from TouchWiz) Samsung Experience 9.x (based on Android 8 "Oreo") (Minor update) Samsung Experience 10.x (based on Android 9 "Pie) (Minor and Last update before redesign One UI) Current One UI version list One UI 1.x (based on Android 9 "Pie") (Initial release) One UI 2.x (based on Android 10) (Minor UI update) One UI 3.x (based on Android 11) (Minor UI update) One UI 4.x (based on Android 12) (Upcoming update) OxygenOS OxygenOS is based on the open source Android Open Source Project (AOSP) and is developed by OnePlus to replace Cyanogen OS on OnePlus devices such as the OnePlus One, and it is preinstalled on the OnePlus 2, OnePlus X, OnePlus 3, OnePlus 3T, OnePlus 5, OnePlus 5T and OnePlus 6. As stated by Oneplus, OxygenOS is focused on stabilizing and maintaining of stock like those found on Nexus devices. It consists of mainly Google apps and minor UI customization to maintain the sleekness of pure Android. Current OxygenOS version list Oxygen OS 1.0.x (based on Android 5.0.x "Lollipop") (initial release) Oxygen OS 2.0.x (based on Android 5.1.x "Lollipop") (overall maintenance update) Oxygen OS 3.0.x (based on Android 6.0 "Marshmallow") (major Android update) Oxygen OS 3.1.x (based on Android 6.0.1 "Marshmallow") (minor maintenance update) Oxygen OS 3.2.x (based on Android 6.0.1 "Marshmallow") (major Android update) Oxygen OS 4.x.x (based on Android 7.x "Nougat") (major Android update) Oxygen OS 5.x.x (based on Android 8.x "Oreo") (major Android update) Oxygen OS 9.x.x (based on Android 9.0 "Pie") (major Android update) Oxygen OS 10.x.x (based on Android 10.0 "10") (major Android update) Oxygen OS 11.x.x (based on Android 11.0 "11") (major Android update) Oxygen OS 12.x.x (based on Android 12.0 "12") (major Android update) Pixel UI (Pixel Launcher) Google Pixel UI or Pixel Launcher is developed by Google that is based on open source Android. Unlike the Nexus phones, where Google shipped with the "stock" Android, the UI that came with the first generation Pixel phones were slightly modified as compared to the "stock" Android. As part of the Google Pixel software, the Pixel UI and its home launcher are closed source and proprietary, thus it is only available on the Pixel family devices. (However, third party mods allow non Pixel smartphones to install Pixel Launcher with Google Now feed integration). Current Google Pixel Launcher version list Pixel Launcher – "7.1.1" (based on Android 7.x "Nougat") (Initial release) Pixel Launcher – "8.1.0" (based on Android 8.x "Oreo") (Minor UI update) Pixel Launcher – "9.0" (based on Android 9.0 "Pie") (Minor UI update) Pixel Launcher – "10.0" (based on Android 10.0 "10") (Moderate UI update that support themes) Pixel Launcher – "11.0" (based on Android 11.0 "11") (Minor UI update) Pixel Launcher - "12.0" (based on Android 12.0 "12") (Major UI update) realme UI realme UI is a mobile operating system developed by Realme which is based on OPPO ColorOS, which itself is based on the Android Open Source Project (AOSP). The overall UI mostly resemble its predecessor, however with their own customized UI and element on top of the ColorOS to match their target audience. Current realme UI version list realme UI 1.0 - Based on ColorOS 7.0 - Android 10 - Initial Release realme UI 2.0 - Based on ColorOS 11.0 - Android 11 realme UI 3.0 - Based on ColorOS 12.0 - Android 12 Replicant OS Replicant is a custom mobile operating system based on the Android with all proprietary drivers and bloat closed source software removed. TCL UI TCL UI is a custom user interface that is developed by TCL Technology for their in house smartphone series, the OS is based on the Android Open Source Project (AOSP). Current TCL UI version list TCL UI 1.x - Based on Android 9 "Pie" and Android 10 - Initial Release Xperia UI Sony Xperia UI (formerly known as Sony Ericsson Timescape UI) was the front-end UI that developed by Sony Mobile (formerly Sony Ericsson) In 2010 for their Sony Xperia series that is running Android platform. Sony Xperia UI mostly consists of Sony own's application such as Sony Music (formerly known as Walkman Music player), Albums and Video Player. During its time as Timescape UI, the UI was different than the standard Android UI, instead of traditional apps dock on the bottom part, it was located at the four corner of the home screen, while the middle of the screen consist of the widget. However the recent development of UI are closely resemble to those of Stock Android. Current Xperia UI version list: Timescape version 1 – based on Android "Eclair" 2.0/2.1, initial release Timescape version 2 – based on Android "Gingerbread" 2.3.x, redesigned UI Xperia UI version 3 – based on Android "Gingerbread" and "Ice Cream Sandwich" 2.3.x and 4.0.x, redesigned UI Xperia UI version 4 – based on Android "Jelly Bean" 4.2.x – 4.3.x, redesigned UI Xperia UI version 5 – based on Android "KitKat" 4.4.x, redesigned UI Xperia UI version 6 – based on Android "Lollipop" 5.0.x – 5.1.x, redesigned UI Xperia UI version 7 – based on Android "Marshmallow" 6.0.x, redesigned UI Xperia UI version 8 – based on Android "Nougat" 7.x, redesigned UI Xperia UI version 9 – based on Android "Oreo" 8.x, redesigned UI ZenUI ZenUI is a front-end touch interface developed by ASUS with partners, featuring a full touch user interface. ZenUI is used by Asus for its Android phones and tablet computers, and is not available for licensing by external parties. ZenUI also comes preloaded with Asus-made apps like ZenLink (PC Link, Share Link, Party Link & Remote Link). Current ZenUI version list: ZenUI 1.0 – based on Android "Jelly Bean" and "KitKat" 4.3.x and 4.4.x, initial release ZenUI 2.0 – based on Android "Lollipop" 5.0.x – 5.1.x, redesigned UI ZenUI 3.0 – based on Android "Marshmallow" 6.0.x, redesigned UI ZenUI 4.0 – based on Android "Nougat" 7.x, redesigned UI ZenUI 5.0 – based on Android "Oreo" 8.x, redesigned UI ZenUI 6.0 – based on Android "Pie" 9.0, redesigned UI ZenUI 7.0 – based on Android 10, redesigned UI ZenUI 8.0 – based on Android 11, redesigned UI ZUI ZUI is a custom operating that is original develop by Lenovo subsidiary ZUK Mobile for their smartphones, however after the shutting down of ZUK Mobile, Lenovo had taken over as the main developer to continue developing the ZUI. The operating system is based on the Android Open Source Project (AOSP). ZUI 1.x - Initial Release ZUI 2.x ZUI 3.x ZUI 4.x ZUI 4.x ZUI 10.x (Based on Android 9 "Pie") ZUI 11.x (Based on Android 9 "Pie" and Android 10) Wear OS Wear OS (also known simply as Wear and formerly Android Wear) is a version of Google's Android operating system designed for smartwatches and other wearables. By pairing with mobile phones running Android version 6.0 or newer, or iOS version 10.0 or newer with limited support from Google's pairing application, Wear OS integrates Google Assistant technology and mobile notifications into a smartwatch form factor. In May 2021 at Google I/O, Google announced a major update to the platform, internally known as Wear OS 3.0. It incorporates a new visual design inspired by Android 12, and Fitbit exercise tracking features. Google also announced a partnership with Samsung Electronics, who is collaborating with Google to unify its Tizen-based smartwatch platform with Wear OS, and has committed to using Wear OS on its future smartwatch products. The underlying codebase was also upgraded to Android 11. Wear OS 3.0 will be available to Wear OS devices running Qualcomm Snapdragon Wear 4100 system on chip, and will be an opt-in upgrade requiring a factory reset to install. Current Wear OS version list: Android Wear 4.4w (Based on Android 4.4 "KitKat") - (Initially release) Android Wear 1.0 - 1.3 (Based on Android 5.0 "Lollipop) - (Minor update) Android Wear 1.4 (Based on Android 6.0 "Marshmellow) - (Minor update) Android Wear 2.0 - 2.6 (7.1.1W2) (Based on Android 7.1 "Nougat") - (Minor update) Android Wear 2.6 (7.1.1W3, 8.0.0 W1) - 2.9 (7.1.1W6, 8.0.0W4) (Baded on Android 8.0 "Oreo") - (Minor update) Wear OS 1.0 (Based on Android 8.0 "Oreo") - (Renamed and Minor update) Wear OS 2.0 (Based on Android 8.0 "Oreo") - (Minor update) Wear OS 2.2 (Based on Android 9.0 "Pie") - (Minor update) Wear OS 3.0 (Based on Android 11) - (Major UI and system update) One UI Watch One UI Watch is the user interface Samsung developed for their Wear OS based smartwatch, officially announced after both Google and Samsung confirm both company is going to unify their wearable operating system (Google Wear OS 2.0 and Samsung Tizen) into Wear OS 3.0. Current One UI Watch version list: One UI Watch 3.0 (Based on Wear OS 3.0 - Android 11)(Initially release) Chrome OS Chrome OS is an operating system designed by Google that is based on the Linux kernel and uses the Google Chrome web browser as its principal user interface. As a result, Chrome OS primarily supports web applications. Google announced the project in July 2009, conceiving it as an operating system in which both applications and user data reside in the cloud: hence Chrome OS primarily runs web applications. Due to increase of popularity with 2-in-1 PCs, most recent Chromebooks are introduced with touch screen capability, with Android applications starting to become available for the operating system in 2014. And in 2016, access to Android apps in the entire Google Play Store was introduced on supported Chrome OS devices. With the support of Android applications, there are Chromebook devices that are positioned as tablet based instead of notebooks. Chrome OS is only available pre-installed on hardware from Google manufacturing partners. An open source equivalent, Chromium OS, can be compiled from downloaded source code. Early on, Google provided design goals for Chrome OS, but has not otherwise released a technical description. Sailfish OS Sailfish OS is from Jolla. It is open source with GNU General Public License (GPL) for middleware stack core which comes from MER. Sailfish due to Jolla's business model and due to alliances with various partners and due to intentional design of OS internals, is capable to adopt in several layers third-party software including Jolla software e.g. Jolla's UI is proprietary software (closed source), so such components can be proprietary with many different kinds of licences. However, user can replace them with open source components like e.g. NEMO UI instead Jolla's UI. After Nokia abandoned in 2011 the MeeGo project, most of the MeeGo team left Nokia, and established Jolla as a company to use MeeGo and Mer business opportunities. The MER standard allows it to be launched on any hardware with kernel compatible with MER. In 2012, Linux Sailfish OS based on MeeGo and using middleware of MER core stack distribution was launched for public use. The first device, the Jolla smartphone, was unveiled on May 20, 2013. In 2015, Jolla Tablet was launched and the BRICS countries declared it an officially supported OS there. Jolla started licensing Sailfish OS 2.0 for third parties. Some devices sold are updateable to Sailfish 2.0 with no limits. Nemo Mobile is a community-driven OS, similar to Sailfish but attempting to replace its proprietary components, such as the user interface. Each Sailfish OS version release is named after a Finnish lake: Tizen Tizen (based on the Linux kernel) is a mobile operating system hosted by Linux Foundation, together with support from the Tizen Association, guided by a Technical Steering Group composed of Intel and Samsung. Tizen is an operating system for devices including smartphones, tablets, In-Vehicle Infotainment (IVI) devices, however currently it mainly focus on wearable and smart TVs. It is an open source system (however the SDK was closed source and proprietary) that aims to offer a consistent user experience across devices. Tizen's main components are the Linux kernel and the WebKit runtime. According to Intel, Tizen "combines the best of LiMo and MeeGo." HTML5 apps are emphasized, with MeeGo encouraging its members to transition to Tizen, stating that the "future belongs to HTML5-based applications, outside of a relatively small percentage of apps, and we are firmly convinced that our investment needs to shift toward HTML5." Tizen will be targeted at a variety of platforms such as handsets, touch pc, smart TVs and in-vehicle entertainment. On May 17, 2013, Tizen released version 2.1, code-named Nectarine. While Tizen itself was open source, most of the UX and UI layer that developed by Samsung was mainly closed source and proprietary, such as the TouchWiz UI on the Samsung Z's series smartphone and One UI for their Galaxy Watch wearable lines. KaiOS KaiOS is from Kai. It is based on Firefox OS/Boot to Gecko. Unlike most mobile operating systems which focus on smartphones, KaiOS was developed mainly for feature phones, giving these access to more advanced technologies usually found on smartphones, such as app stores and Wi-Fi/4G capabilities. It is a mix of closed-source and open-source components. FirefoxOS/B2G was released under the permissive MPL 2.0. It does not redistribute itself under the same license, so KaiOS is now presumably proprietary (but still mostly open-source, publishing its source code). KaiOS is not entirely proprietary, as it uses the copyleft GPL Linux kernel also used in Android. Fully open-source, entirely permissive licenses Fuchsia Fuchsia is a capability-based, real-time operating system (RTOS) currently being developed by Google. It was first discovered as a mysterious code post on GitHub in August 2016, without any official announcement. In contrast to prior Google-developed operating systems such as Chrome OS and Android, which are based on Linux kernels, Fuchsia is based on a new microkernel called "Zircon", derived from "Little Kernel", a small operating system intended for embedded systems. This allows it to remove Linux and the copyleft GPL underwhich the Linux kernel is licensed; Fuchsia is licensed under the permissive BSD 3-clause, Apache 2.0, and MIT licenses. Upon inspection, media outlets noted that the code post on GitHub suggested Fuchsia's capability to run on universal devices, from embedded systems to smartphones, tablets and personal computers. In May 2017, Fuchsia was updated with a user interface, along with a developer writing that the project was not a for experimental, prompting media speculation about Google's intentions with the operating system, including the possibility of it replacing Android. LiteOS LiteOS is a lightweight open source real-time operating system which is part of Huawei's "1+2+1" Internet of Things solution, which is similar to Google Android Things and Samsung Tizen. It is released under the permissive BSD 3-clause license. Huawei LiteOS features lightweight, low-power, fast-response, multi-sensor collaboration, multi-protocol interconnect connectivity, enabling IoT terminals to quickly access the network. Huawei LiteOS will make intelligent hardware development easier. Thereby accelerating the realization of the interconnection of all things. Currently LiteOS are introduce to the consumer market with the Huawei Watch GT series and their sub-brand Honor Magic Watch series. Fully open-source, mixed copyleft and permissive licenses PureOS PureOS is a Debian GNU/Linux derivative using only free software meeting the Debian Free Software Guidelines, mainly the copyleft GPL. PureOS is endorsed by Free Software Foundation as one of the freedom-respecting operating systems. It is developed by Purism, and was already in use on Purism's laptops before it was used on the Librem 5 smartphone. Purism, in partnership with GNOME and KDE, aims to separate the CPU from the baseband processor and include hardware kill switches for the phone's Wi-Fi, Bluetooth, camera, microphone and baseband processor, and provide both GNOME and KDE Plasma Mobile as options for the desktop environment. Ubuntu Touch Ubuntu Touch is an open-source (GPL) mobile version of the Ubuntu operating system originally developed in 2013 by Canonical Ltd. and continued by the non-profit UBports Foundation in 2017. Ubuntu Touch can run on a pure GNU/Linux base on phones with the required drivers, such as the Librem 5 and the PinePhone. To enable hardware that was originally shipped with Android, Ubuntu Touch makes use of the Android Linux kernel, using Android drivers and services via an LXC container, but does not use any of the Java-like code of Android. As of February 2022, Ubuntu Touch is available on 78 different phones and tablets. The UBports Installer serves as an easy-to-use tool to allow inexperienced users to install the operating system on third-party devices without damaging their hardware. Plasma mobile Plasma Mobile is a Plasma variant for smartphones. Plasma Mobile runs on Wayland and it is compatible with Ubuntu Touch applications, PureOS applications, and eventually Android applications via KDE's Shashlik project also sponsored by Blue Systems, or Anbox. It is under the copyleft GPLv2 license. The Necuno phone uses Plasma Mobile. It is which is entirely open-source and thus does not have a cellular modem, so it must make calls by VOIP, like a pocket computer. PostmarketOS PostmarketOS is based on the Alpine Linux Linux distribution. It is intended to run on older phone hardware. it is in alpha. LuneOS LuneOS is a modern reimplementation of the Palm/HP webOS interface. Closed source iOS iOS (formerly named iPhone OS) was created by Apple Inc. It has the second largest installed base worldwide on smartphones, but the largest profits, due to aggressive price competition between Android-based manufacturers. It is closed source and proprietary, and is built on the open source Darwin operating system. The iPhone, iPod Touch, iPad and second or third-generation Apple TV all use iOS, which is derived from macOS. Native third-party applications were not officially supported until the release of iPhone OS 2.0 on July 11, 2008. Before this, "jailbreaking" allowed third-party applications to be installed. In recent years, the jailbreaking scene has changed drastically due to Apple's continued efforts to secure their operating system and prevent unauthorized modifications. Currently, jailbreaks of recent iterations of iOS are only semi-untethered, which requires a device to be re-jailbroken at every boot, and exploits for jailbreaks are becoming increasingly hard to find and use. Currently all iOS devices are developed by Apple and manufactured by Foxconn or another of Apple's partners. iPadOS iPadOS is a tablet operating system created and developed by Apple Inc. specifically for their iPad line of tablet computers. It was announced at the company's 2019 Worldwide Developers Conference (WWDC), as a derivation from iOS but with a greater emphasis put on multitasking. It was released on September 24, 2019. watchOS watchOS is the operating system of the Apple Watch, developed by Apple Inc. It is based on the iOS operating system and has many similar features. It was released on April 24, 2015, along with the Apple Watch, the only device that runs watchOS. It is currently the most widely used wearable operating system. Its features focus on convenience, such as being able to place phone calls and send texts, and health, such as fitness and heart rate tracking. The most current version of the watchOS operating system is watchOS 7. Kindle firmware Kindle firmware is mobile operating system specifically design for Amazon Amazon Kindle e-readers, it is based on custom Linux kernel however it is entirely closed source and proprietary, and only runs on Amazon Kindle line up manufactured under Amazon brand. Windows 10 Windows 10 (not to be confused with Windows 10 Mobile—see below) is a personal computer operating system developed and released by Microsoft as part of the Windows NT family of operating systems. It was released on July 29, 2015, and many editions and versions have been released since then. Just like its predecessors, it was designed to run across multiple Microsoft product such as PCs and Tablets. The Windows user interface was revised to handle transitions between a mouse-oriented interface and a touchscreen-optimized interface based on available input devices‍—‌particularly on 2-in-1 PCs. Windows 10 also introduces the universal apps, expanding on Metro-style apps, these apps can be designed to run across multiple Microsoft product families with nearly identical code‍—‌including PCs, tablets, smartphones, embedded systems, Xbox One, Surface Hub and Mixed Reality. Windows 11 Windows 11 is a major version of the Windows NT operating system developed by Microsoft that was announced on June 24, 2021, and is the successor to Windows 10, which was released n 2015. Windows 11 was released on October 5, 2021, as a free upgrade via Windows Update for eligible devices running Windows 10. Microsoft promoted that Windows 11 would have improved performance and ease of use over Windows 10; it features major changes to the Windows shell influenced by the canceled Windows 10X, including a redesigned Start menu, the replacement of its "live tiles" with a separate "Widgets" panel on the taskbar, the ability to create tiled sets of windows that can be minimized and restored from the taskbar as a group, and new gaming technologies inherited from Xbox Series X and Series S such as Auto HDR and DirectStorage on compatible hardware. Internet Explorer is fully replaced by the Blink layout engine-based Microsoft Edge, while Microsoft Teams is integrated into the Windows shell. Microsoft also announced plans to offer support for Android apps to run on Windows 11, with support for Amazon Appstore and manually-installed packages. Similar to Windows 10, it was designed to run across multiple Microsoft product such as PCs and Tablets. The Windows user interface was further revised to combine the UI element of both mouse-oriented interface and a touchscreen-optimized interface based into a hybrid UI that combined touch and traditional desktop UI. Minor proprietary operating system Other than the major mobile operating system from the major tech company, some company such as Huami(Amazfit), Huawei, realme, TCL and Xiaomi has their own proprietary RTOS that develop specifically for their own smartband and smartwatch that is design to be power efficient and low battery consumption that is neither based on Android nor Linux Kernel. Proprietary Amazfit OS Operating System that is primarily design for their Bip series, however Huami is currently developing the operating system to run on other company smartwatch as well. (Not to be confused with the Android-based Amazfit OS as both of them sharing the name yet both are based on different operating system, for their Android based Amazfit OS, kindly refer to the Android section) Huawei/Honor Band Operating System Huawei Band Operating system is an operating system specifically designed and developed by Huawei for their fitness tracker, including smartband series from Honor. (Not to be confused with another RTOS (LiteOS) which was also developed by Huawei.) Lenovo RTOS Proprietary OS develop by Lenovo for their fitness tracker and smartwatch. realme Wearable Operating System A proprietary operating system design to run on realme smartband and smartwatch. TCL Wearable Real Time Operating System A proprietary RTOS powering TCL and Alcatel brand smartband and smartwatch. Xiaomi Mi Band Operating System Proprietary RTOS that is develop by Huami for Xiaomi Mi Band series. (Not to be confused with Xiaomi MIUI for smartwatch which is based on Wear OS) Under maintenance only BlackBerry 10 BlackBerry 10 (based on the QNX OS) is from BlackBerry. As a smartphone OS, it is closed source and proprietary, and only runs on phones and tablets manufactured by BlackBerry. One of the dominant platforms in the world in the late 2000s, its global market share was reduced significantly by the mid-2010s. In late 2016, BlackBerry announced that it will continue to support the OS, with a promise to release 10.3.3. Therefore, BlackBerry 10 would not receive any major updates as BlackBerry and its partners would focus more on their Android base development. Discontinued software platforms Open source CyanogenMod CyanogenMod was a custom mobile operating system based on the Android Open Source Project (AOSP). It was a custom ROM that was co-developed by the CyanogenMod community. The OS did not include any proprietary apps unless the user installed them. Due to its open source nature, CyanogenMod allowed Android users who could no longer obtain update support from their manufacturer to continue updating their OS version to the latest one based on official releases from Google AOSP and heavy theme customization. The last version of the OS was CyanogenMod 13 which was based on Android Asus On December 24, 2016, CyanogenMod announced on their blog that they would no longer be releasing any CyanogenMod updates. All development moved to LineageOS. Cyanogen OS Cyanogen OS was based on CyanogenMod and maintained by Cyanogen Inc; however, it included proprietary apps and it was only available for commercial uses. Firefox OS Firefox OS (formerly known as "Boot to Gecko" and shortly "B2G") is from Mozilla. It was an open source mobile operating system released under the Mozilla Public License built on the Android Linux kernel and used Android drivers, but did not use any Java-like code of Android. According to Ars Technica, "Mozilla says that B2G is motivated by a desire to demonstrate that the standards-based open Web has the potential to be a competitive alternative to the existing single-vendor application development stacks offered by the dominant mobile operating systems." In September 2016, Mozilla announced that work on Firefox OS has ceased, and all B2G-related code would be removed from mozilla-central. MeeGo/Maemo/Moblin MeeGo was from non-profit organization The Linux Foundation. It is open source and GPL. At the 2010 Mobile World Congress in Barcelona, Nokia and Intel both unveiled MeeGo, a mobile operating system that combined Moblin and Maemo to create an open-sourced experience for users across all devices. In 2011 Nokia announced that it would no longer pursue MeeGo in favor of Windows Phone. Nokia announced the Nokia N9 on June 21, 2011, at the Nokia Connection event in Singapore. LG announced its support for the platform. Maemo was a platform developed by Nokia for smartphones and Internet tablets. It is open source and GPL, based on Debian GNU/Linux and draws much of its graphical user interface (GUI), frameworks, and libraries from the GNOME project. It uses the Matchbox window manager and the GTK-based Hildon as its GUI and application framework. webOS webOS was developed by Palm. webOS is an open source mobile operating system running on the Linux kernel, initially developed by Palm, which launched with the Palm Pre. After being acquired by HP, two phones (the Veer and the Pre 3) and a tablet (the TouchPad) running webOS were introduced in 2011. On August 18, 2011, HP announced that webOS hardware would be discontinued, but would continue to support and update webOS software and develop the webOS ecosystem. HP released webOS as open source under the name Open webOS, and plans to update it with additional features. On February 25, 2013, HP announced the sale of webOS to LG Electronics, who used the operating system for its "smart" or Internet-connected TVs. However, HP retained patents underlying WebOS and cloud-based services such as the App Catalog. Closed source BlackBerry OS In 1999, Research In Motion released its first BlackBerry devices, providing secure real-time push-email communications on wireless devices. Services such as BlackBerry Messenger provide the integration of all communications into a single inbox. In September 2012, RIM announced that the 200 millionth BlackBerry smartphone was shipped. As of September 2014, there were around 46 million active BlackBerry service subscribers. In the early 2010s, RIM has undergone a platform transition, changing its company name to BlackBerry Limited and making new devices on a new platform named "BlackBerry 10". Symbian Symbian platform was developed by Nokia for some models of smartphones. It is proprietary software, it was however used by Ericsson (SonyEricsson), Sending and Benq. The operating system was discontinued in 2012, although a slimmed-down version for basic phones was still developed until July 2014. Microsoft officially shelved the platform in favor of Windows Phone after its acquisition of Nokia. Bada Bada platform (stylized as bada; Korean: 바다) was an operating system for mobile devices such as smartphones and tablet computers. It was developed by Samsung Electronics. Its name is derived from "바다 (bada)", meaning "ocean" or "sea" in Korean. It ranges from mid- to high-end smartphones. To foster adoption of Bada OS, since 2011 Samsung reportedly has considered releasing the source code under an open-source license, and expanding device support to include Smart TVs. Samsung announced in June 2012 intentions to merge Bada into the Tizen project, but would meanwhile use its own Bada operating system, in parallel with Google Android OS and Microsoft Windows Phone, for its smartphones. All Bada-powered devices are branded under the Wave name, but not all of Samsung's Android-powered devices are branded under the name Galaxy. On February 25, 2013, Samsung announced that it will stop developing Bada, moving development to Tizen instead. Bug reporting was finally terminated in April 2014. Palm OS Palm OS/Garnet OS was from Access Co. It is closed source and proprietary. webOS was introduced by Palm in January 2009, as the successor to Palm OS with Web 2.0 technologies, open architecture and multitasking abilities. Microsoft Windows Mobile Windows Mobile is a family of proprietary operating systems from Microsoft aimed at business and enterprise users, based on Windows CE and originally developed for Pocket PC (PDA) devices. In 2010 it was replaced with the consumer-focused Windows Phone. Versions of Windows Mobile came in multiple editions, like "Pocket PC Premium," "Pocket PC Professional," "Pocket PC Phone," and "Smartphone" (Windows Mobile 2003) or "Professional," "Standard," and "Classic" (Windows Mobile 6.0). Some editions were touchscreen-only and some were keyboard-only, although there were cases where device vendors managed to graft support for one onto an edition targeted at the other. Cellular phone features were also only supported by some editions. Microsoft started work on a version of Windows Mobile that would combine all features together, but it was aborted, and instead they focused on developing the non-backwards-compatible, touchscreen-only Windows Phone 7. Windows Phone Windows Phone is a proprietary mobile operating system developed by Microsoft for smartphones as the replacement successor to Windows Mobile and Zune. Windows Phone features a new touchscreen-oriented user interface derived from Metro design language. Windows Phone was replaced by Windows 10 Mobile in 2015. Windows 10 Mobile Windows 10 Mobile (formerly called Windows Phone) is from Microsoft. It is closed source and proprietary. Unveiled on February 15, 2010, Windows Phone includes a user interface inspired by Microsoft's Metro Design Language. It is integrated with Microsoft services such as OneDrive and Office, Xbox Music, Xbox Video, Xbox Live games and Bing, but also integrates with many other non-Microsoft services such as Facebook and Google accounts. Windows Phone devices were made primarily by Microsoft Mobile/Nokia, and also by HTC and Samsung. On January 21, 2015, Microsoft announced that the Windows Phone brand will be phased out and replaced with Windows 10 Mobile, bringing tighter integration and unification with its PC counterpart Windows 10, and provide a platform for smartphones and tablets with screen sizes under 8 inches. On October 8, 2017, Microsoft officially announced that they would no longer push any major updates to Windows 10 Mobile, instead it would put it in maintenance mode, where Microsoft would push bug fixes and general improvements only, therefore Windows 10 Mobile would not receive any new feature updates. On January 18, 2019, Microsoft announced that support for Windows 10 Mobile would end on December 10, 2019, with no further security updates released after then, and that Windows 10 Mobile users should migrate to iOS or Android phones. Current Windows 10 Mobile version list: Windows 10 Mobile – Version 1511 (November Update "Threshold") – major UI update Windows 10 Mobile – Version 1607 (Anniversary Update "Redstone 1") Windows 10 Mobile – Version 1703 (Creators Update "Redstone 2") Windows 10 Mobile – Version 1709 (Fall Creators Update) Market share Usage In 2006, Android and iOS did not exist and only 64 million smartphones were sold. In 2018 Q1, 383.5 million smartphones were sold and global market share was 85.9% for Android and 14.1% for iOS. Other OS smartphones were 0,131 million, counting the 0.03% of the total. According to StatCounter web use statistics (a proxy for all use), smartphones (alone without tablets) have majority use globally, with desktop computers used much less (and Android in particular more popular than Windows). Use varies however by continent with smartphones way more popular in the biggest continents, i.e. Asia, and the desktop still more popular in some, though not in North America. The desktop is still popular in many countries (while overall down to 44.9% in the first quarter of 2017), smartphones are more popular even in many developed countries (or about to be in more). A few countries on any continent are desktop-minority; European countries (and some in South America, and a few, e.g. Haiti, in North America; and most in Asia and Africa) are smartphone-majority, Poland and Turkey highest with 57.68% and 62.33%, respectively. In Ireland, smartphone use at 45.55% outnumbers desktop use and mobile as a whole gains majority when including the tablet share at 9.12%. Spain is also slightly desktop-minority. The range of measured mobile web use varies a lot by country, and a StatCounter press release recognizes "India amongst world leaders in use of mobile to surf the internet" (of the big countries) where the share is around (or over) 80% and desktop is at 19.56%, with Russia trailing with 17.8% mobile use (and desktop the rest). Smartphones (alone, without tablets), first gained majority in December 2016 (desktop-majority was lost the month before), and it wasn't a Christmas-time fluke, as while close to majority after smartphone majority happened again in March 2017. In the week from November 7–13, 2016, smartphones alone (without tablets) overtook desktop, for the first time (for a short period; non-full-month). Mobile-majority applies to countries such as Paraguay in South America, Poland in Europe and Turkey; and most of Asia and Africa. Some of the world is still desktop-majority, with e.g. in the United States at 54.89% (but no not on all days). However, in some territories of the United States, such as Puerto Rico, desktop is way under majority, with Windows under 30% overtaken by Android. On October 22, 2016 (and subsequent weekends), mobile showed majority. Since October 27, the desktop hasn't shown majority, not even on weekdays. And smartphones alone have showed majority since December 23 to the end of the year, with the share topping at 58.22% on Christmas Day. To the "mobile"-majority share then of smartphones, tablets could be added giving a 63.22% majority. While an unusually high top, a similarly high also happened on Monday April 17, 2017, with then only smartphones share slightly lower and tablet share slightly higher, with them combined at 62.88%. , the world has turned desktop-minority; at about 49% desktop use for the previous month, but mobile wasn't ranked higher, tablet share had to be added to it to exceed desktop share. By now, mobile (smartphones) have full majority, outnumbering desktop/laptop computers by a safe margin (and no longer counting tablets with desktops makes them most popular). By operating system Note: Windows includes all versions. BlackBerry includes all versions. Other includes all other smartphone OSes but not feature phone OSes. See also Comparison of mobile operating systems Usage share of operating systems Real-time operating system List of GPS software for mobile phones Mobile device Smartphone Tablet computer Information appliance List of open-source mobile phones Operating system Network operating system References External links Android Apple Operating systems Mobile phones Software wars
5919849
https://en.wikipedia.org/wiki/Zune
Zune
Zune is a discontinued line of digital media products and services marketed by Microsoft from November 2006 until its discontinuation in June 2012. Zune consisted of a line of portable media players, digital media player software for Windows PCs, a music subscription service known as a "Zune Music Pass", music and video streaming services for the Xbox 360 game console via the Zune Software, music, TV and movie sales, and desktop sync software for Windows Phone. Zune was also the provider of music streaming for United Airlines in-flight, after a partnership in 2010. The Zune hardware players were discontinued in October 2011. In June 2012, Microsoft announced plans to discontinue all "Zune" products, and instead, Microsoft would distribute its digital media content and services under the Xbox Music and Xbox Video names, available on their line of products, including Windows 8 PCs and tablets, Xbox 360 game console, and Windows Phone smart phones. The www.zune.net domain now redirects to Xbox's website, but the software retained the Zune name. The Windows Phone App succeeded Zune Software as the desktop sync service for Windows Phone 8, as part of Microsoft's discontinuation of the Zune brand. However, Zune Software must still be used for Windows Phone 7 desktop sync, and is still available to download from the Windows Phone website for all Windows Phone 7 devices. In November 2015, Microsoft retired the Zune music download and streaming service. Remaining Zune subscribers were switched over to Microsoft's Groove Music platform, whose subscription services closed on December 31, 2017. History MSN Music Zune music and devices were follow-on to Microsoft's MSN Music service. MSN Music was created in 2004 to compete with Apple's iTunes services. After only two years, Microsoft announced the closing of MSN Music in 2006 immediately before announcing the Zune service. In 2008, Microsoft shut down the MSN Music licensing servers only two years after promising users the servers would be available for five years. First generation The first-generation Zune device was created by Microsoft in close cooperation with Toshiba, which took the design of the Gigabeat S and redeveloped it under the name Toshiba 1089 as registered with the Federal Communications Commission (FCC) starting in 2006. Xbox 360 overseer J Allard ran the project, codenamed "Argo", staffed with Xbox and MSN Music Store developers who worked on "Alexandria", finalized as Zune Marketplace. Both products were later united under the Zune brand name in the U.S. market. While some features were praised, the initial Zune has been regarded with derision and jokes for its bulky size and brown color, with CNET regarding it at best as a "high-profile underdog alternative to Apple's iPod" where other Windows Media MP3 players from Creative, iRiver, and Samsung had not succeeded. At midnight on December 31, 2008, many first generation Zune 30 models froze. Microsoft stated that the problem was caused by the internal clock driver written by Freescale and the way the device handles a leap year. It automatically fixed itself 24 hours later, but an intermediate "fix", for those who did not wish to wait, was to drain the device's battery and then recharge after 12 noon GMT on January 1, 2009. The first generation and later Zune devices included a number of social features, including the ability to share songs with other Zune users wirelessly. Songs that had been transferred over wi-fi could then be played three times over three days. Second generation The first wave of the second-generation (Zune 4, 8, and 80, manufactured by Flextronics), introduced the touch-sensitive Zune Pad, which was shaped like a squircle. The 4 and 8 GB Zune devices use flash memory and are smaller in size than the 80 GB version, which uses a hard drive. The 30 GB Zune was not redesigned, although it received a software update that brought its interface in line with the second generation models. At the same time, the Zune 2.0 software was released for Windows PCs. This version of the software was completely re-written and featured a new user interface. Zune devices feature games developed using XNA. An early version of XNA Game Studio 3.0, released in May 2008, allowed developers to work on games for Zune devices. The second and last wave of the second-generation (Zune 16 and 120) released in September 2008, coinciding with the release of the Zune Software 3.0 update. Included in this firmware update was the ability to tag and later purchase songs heard on FM radio, channels which can be customized to deliver suggested songs for the user, the games Hexic and Texas Hold' em, support for audiobooks from online stores such as Audible.com and others that support OverDrive media files, a clock, and changed quicklist functionality. The ability to purchase songs from Zune Marketplace on the device while connected to the Internet via Wi-Fi was also introduced. To help push this feature, Microsoft partnered with Wayport to allow Zune devices to access its network of over 10,000 wireless hotspots, including those at McDonald's restaurants. Zune Pass customers in the United States could also now download 10 tracks to keep per month in addition to the existing subscription-dependent unlimited music downloads. Third generation The first wave of the third-generation (Zune HD 16 GB and 32 GB) released in September 2009. On the same day, the 4.0 software update of the desktop software was released to support the new devices. In addition, it became possible for Zune Pass subscribers to stream tracks through a computer's web browser. Zune 4.0 also supports internet radio streams but this feature is disabled by default and can only be enabled by a third-party patch. This device release marked the first time Zune firmware was released for the Zune line that did not provide new features for older models. These models were given a firmware update as version 3.2. The second and last wave of the third-generation (Zune HD 64 GB) released in April 2010. The same day, desktop and firmware updates introduced SmartDJ and codec features. The firmware update brought picks and improved the TV-out experience on the Zune HD. From Summer 2010, United Airlines started to offer Zune in-flight audio by means of 21 playlists that are very similar to the Zune Channels offered on the Zune Marketplace. Each channel offers up to 3 hours of unique programming ranging from classic rock, contemporary pop, opera, electronica, piano jazz, and others. Discontinuation of Zune hardware On March 15, 2011, Microsoft announced that no new Zune hardware players would be developed, although existing models would remain for sale. The Zune had failed to capture significant market share after five years against the Apple iPod, and a recent study by NPD Group indicated that the Zune did not make the list of the five best-selling portable music players in the U.S. The iPod has been more successful because of simplicity and had better ratings. It has also been suggested that there is a much larger trend that standalone music players were giving way to smartphones with personalized, app-driven music. On October 3, 2011, Microsoft announced the discontinuation of all Zune hardware, encouraging users to transition to Windows Phone. Later, the announcement was removed and a Zune Support Team member tweeted that the page was added to the website in error. Finally, despite previous denials, the original announcement of the Zune hardware's discontinuation was restored to the Zune Support site. On other Microsoft products Xbox 360 In Autumn 2009, movies and TV shows became available through streaming or download through Zune Video Marketplace on Xbox 360. On November 4, 2010, the music portion of the Zune Marketplace was brought to Xbox. This coincided with the launch of the Kinect and Kinect owners can navigate the application menus using hand gestures, without a controller. Users must have a Zune Pass subscription to play music in the application, and only Zune Pass content is available. Locally saved music must still be played through the Xbox's native media library. Microsoft Mobile Phones Starting with the Microsoft KIN ONE and KIN TWO mid 2010, the Zune music services and features from the Zune HD became available on Microsoft's mobile phones. Shortly following the failure of the KIN line, Windows Phone 7 released and included the same Zune music app that was present on KIN. December of that same year, the rerelease of the KIN phones (whose names were denoted simply with "m" as ONEm and TWOm) yet again included the Zune music app, though the phones themselves were now feature-phones instead of smartphones. Each of these first generation Windows Phone smartphones were supported by the Zune desktop software in addition to the Zune devices. Microsoft announced and released new versions of the Zune devices once a year until 2010. In March 2011, Bloomberg.com published an article claiming that Microsoft would stop introducing new versions of the Zune music and video player. The article has been widely quoted over the Internet and by news agencies. However, a Microsoft representative for Zune business development denied this rumor saying that the Windows Phone platform introduction should be considered to be the annual Zune update for 2010. All Windows Phone devices include capacitive multi-touch screens, FM radios, Wi-Fi, and certain other features included on the Zune HD. The user interface of the Zune devices, particularly the Zune HD, served as the inspiration for the user interface of Windows Phone. Microsoft refers to the design language of this user interface as Metro and continued to use it in their products moving forward from Zune HD. (Products with elements of Metro: Microsoft Band & Band 2, KIN, WP7, WP8, Windows 8, W10M, Windows 10, Xbox 360, Xbox One and S/X, and Xbox Series S/X.) On October 11, 2010, Microsoft released Zune software v4.7, which supports syncing of Windows Phone 7 devices with a Windows PC. Zune software was then succeeded as a desktop sync app by the Windows Phone App for Windows Phone 8. Zune devices The first Zune model, the Zune 30, was released in the USA on November 14, 2006, featuring a capacity of 30 gigabytes, FM radio, and a 3-inch screen. The Zune 30 was initially available in black, brown and white. Retail packages contained a pair of basic earbuds, a USB sync cable, a small drawstring pouch, a start-up CD, documentation and a 14-day free trial of the premium Zune Pass streaming service. The Zune 80 was announced on October 2, 2007, along with the smaller Zune 4 and Zune 8 to compete with Apple's iPod nano line. These were to be known as the second generation of Zune devices. The Zune 80 featured a 3.2 inch screen, while the Zune 4 and 8 come with a 1.8inch screen. Both come with a new touchpad-style input device ("squircle") and new software. Additional file support for H.264 and MPEG-4 formats was also included, whereas the older Zune 30 requires these formats to be transcoded to WMV prior to sync. The ability to sync wirelessly (automatically if connected to a power supply), podcast support, and an upgraded song-sharing licensing became available on all models. The new software also allowed a Zune device to communicate with other Zune devices to share pictures and songs. A free firmware update added the new software features to the original Zune 30, and was released on November 13, 2007. The Zune 80 came bundled with a USB connection cord and premium headphones. The Zune 4 and 8 come with a USB connection cord and basic headphones. The Zune 30, the original Zune music player, has a 30GB hard drive, 3 inch screen, and a simple directional pad for menu navigation. The second generation of Zune devices includes the Zune 4, 8, 16, 80 and 120. The Zune 4, 8, and 16 are smaller in size and have 4GB, 8GB, and 16GB of flash memory respectively. All second generation Zunes have a Zune Pad instead of the original directional pad that was included on the Zune 30. Microsoft released an upgrade to the software/firmware for all Zune models, including the Zune 30. On May 26, 2009, Microsoft announced the Zune HD, the first touch screen Zune. The Zune HD has HD Radio and the ability to display video in High Definition through a docking station (sold separately). The screen is multi-touch enabled and uses gestures such as swiping and pinching throughout the player. The device comes with 16GB, 32GB, or 64GB of flash memory. The screen is OLED, 3.3inches, and has a 480x272 16:9 resolution. Also included are Wi-Fi, a custom Internet Explorer browser, and an accelerometer. Preloaded content Zune devices featured preloaded audio and video content from various artists, including Wisin & Yandel, BT, The Thermals, Paul Oakenfold, and CSS. Limited edition, promotional, and employee-gift Zune devices also featured unique content of varying kinds (i.e. pictures, videos, music, and/or audio/video podcasts). Specifications Accessories The standard Zune devices come with basic headphones and a proprietary USB data cable. The Zune 30 comes with these items as well as a carrying bag, and the Zune 80 model has upgraded "Zune Premium" headphones. Accessories that were sold separately included, but are not limited to: Charging devices (car adapter, AC wall-socket adapters, external battery) I/O adapters (A/V composite, FM transmitters, headphones, USB data cable) Docks (charging, multimedia large speaker, vertical hands-free assist) Protection (glass screen protection, hardened/cushioning material case protection) Carrying cases (standard issue, armband type, and belt clip) Replacement parts and upgrades (battery, hard drive, LCD, etc.) Among the firms that made Zune accessories were Microsoft, Altec Lansing, Belkin, Digital Lifestyle Outfitters (DLO), Dual Electronics, Griffin Technology, Harman Kardon, JBL, Integrated Mobile Electronics, Jamo International, Klipsch Audio Technologies, Logitech, Monster Cable Products Inc., Speck, Targus, Kicker and VAF Research. Firmware According to Microsoft, the most up-to-date firmware version is 4.5 (114) for the Zune HD, which replaces the original player firmware that ships on the device, 4.0 (356). In the case of the Zune 4, 8, 16, 30, 80, and 120 players, the most current player software version is 3.3, which provides compatibility with Zune 4.2. Version 3.3 was primarily a bug fix release and was released on January 26, 2010. The operating system for the Zune devices is based on the Windows CE kernel for ARM architecture and uses a distribution similar to the Portable Media Center found on the Gigabeat S. Zune's native file compatible formats are: JPEG for images; WMV (Used by Zune Marketplace) MPEG-4 – supported on all models except the Zune 30 device H.264 – supported on all models except the Zune 30 device Avi video (Xvid) support is included on the Zune HD (firmware versions 4.5 and later). MP3 (used by Zune Marketplace) AAC (unprotected) not AAC (.m4a) WMA Pro (2-channel) WMA Standard (used by Zune Marketplace) WMA lossless The graphical user interface (GUI) (called the "twist interface" by Microsoft) has sections for music, videos, pictures, social, radio, podcasts, marketplace, games and settings. It is said to provide "two-dimensional navigation" for scrolling through items with its directional pad. In the music section, users can add songs to a quick playlist without reconnecting to the desktop software. In the picture section, the background can be customized using any image stored on the device (for viewing) as wallpaper. In the radio section, users can receive and play FM radio internally, with North American, Japanese, and European tuning ranges, and display Radio Data System information (usually artist and song) when available. When artist/song information are available, the device can search for the song in the Zune Marketplace for download or purchase. In the social section, users can broadcast the user's profile and recent activity to others nearby. The first updates to the firmware added sharing features (send, community, list nearby Zune users) as described in FCC filings. Firmware 1.1 allowed the device to inherit sharing capabilities described by codename Pyxis. Early firmware releases patched software bugs. About a year later, the much anticipated 2.2 firmware release added support for DVR-MS (Media Center Recorded TV) files, lossless playback, added wireless syncing, and GUI interface improvements. Zune supports the Windows Media DRM digital rights management system, which is not compatible with other DRM systems and is not part of the PlaysForSure platform or program.  Multimedia content is transferred through Media Transfer Protocol (MTP); however, its proprietary MTP extensions ("MTPZ") place an interoperability barrier between the Zune and previous MTP-based software. Zune software The Zune software functions as management software for the device, a full media player application with a library, an interface to the Zune Marketplace, and as a media streaming server. Zune Software is used to sync with all devices with Zune functionality, including the Zune devices, Windows Phone 7, and Microsoft Kin. Zune devices work exclusively with the Zune software and Marketplace. The Zune software organizes the media in its library and allows users to add to the library by ripping from CDs, syncing with a Zune device, and downloading from the Zune Marketplace. The Zune software also allows one to organize song metadata. It can automatically download album art and metadata tag data for content in the library. Any file in one's Collection that has a non-Zune compatible format is automatically transcoded into a compatible format upon syncing the files to a Zune compatible device. This feature still works as of the release of Windows 11, however, in Windows 10 and 11 the ability to convert videos, specifically, is broken. On the PC, the Zune software streams files to other PCs, the Xbox 360, and other compatible devices. The Zune software also connects with the Zune social and keeps track of files swapped with other users. The Zune software runs only on 32-bit Windows XP or 32-bit/64-bit, Windows Vista, Windows 7, and Windows 8. Windows XP Professional x64 Edition is not supported. Zune Software was succeeded by the Windows Phone App as a desktop sync service for Windows Phone 8. Windows Phone 7, Kin, and all previous Zune devices will still use the Zune software, which is still available for download on the Windows Phone website. Zune Marketplace Zune Marketplace successor services As of October 16, 2012, all Zune Marketplace products and services have been replaced by Xbox Music, Xbox Music Pass, Xbox Video, and Windows Phone Store. The Zune Marketplace was an online store that offered music, podcasts, TV shows, movies, music videos, movie trailers and mobile applications. Content can be viewed or purchased on Windows PCs with the Zune software installed, Zune devices, the Xbox 360, the Microsoft Kin phones, or Windows Phone phones. It offers a selection of 14 million songs and the Zune Pass music subscription service. Availability Zune Marketplace was originally only available in the United States. In October 2010, certain Zune Marketplace content became available in additional countries: the United Kingdom, France, Italy, Spain, Germany, Austria, Belgium, Ireland, the Netherlands, Switzerland, Mexico, Canada, Australia, and New Zealand. However, not all content was available in all countries; for example, podcasts and TV shows were not offered at all outside the United States. Sales and marketing Marketing Microsoft launched several campaigns to jump-start the Zune. It had a major campaign to promote Zune with "Music the way it wants to be" as a major theme and "Welcome to the social" as an advertisement tagline. Also, the company enlisted about 200 "Zune-masters" to advertise the device on American college campuses, to promote the item, and to run Zune-related events. In exchange, they received free merchandise, including a Zune. Additionally, Microsoft launched an attempt at viral marketing with its comingzune site, complete with several videos in succession. Along with ZuneInsider, and several other ad hoc events, Microsoft hoped to generate buzz for the product outside of the normal marketing avenues, and market its product as a part of a social construct. The choice of branding and distribution were part of the Zune as a decision of "two strategies in the market right now: cross-brand ecosystems... and singular brand ecosystems... The former is gaining in share and units sold, but the latter has enormous share and won't give that up easily." Microsoft normally follows a platform (cross-brand) strategy, as exemplified by the PlaysForSure system. However, its Xbox division has gained some experience with the vertically integrated strategy in which it controls everything end-to-end from the hardware to the online store. With Apple dominating the audio market with its vertically integrated iPod system, the Xbox division won permission to try the same approach, separately from PlaysForSure and PlayReady. Microsoft also wanted to go beyond Apple's efforts and promote the tagline "the social" and wireless sharing as key differentiators. Chris Stephenson, leader of Zune's marketing and manager of Global Marketing for the Entertainment Business, said, "we see a great opportunity to bring together technology and community to allow consumers to explore and discover music together." New York Times Magazine columnist Rob Walker agrees that the Zune's "community and togetherness seem like a reasonable counterpunch to iPod's supposed attraction as an individuality enabler that allows owners to wallow in their own tasteful personal soundtracks." But he also sees the Zune as having gained appeal as an individualistic statement against the omnipresent iPod: "The most salient feature of the Zune seems to be that it's not an iPod". Microsoft also released a Zune theme for Windows XP that replaced the appearance of the operating system. This theme includes an orange Start button and black taskbar/Start menu. Zune has also expanded its brand efforts by creating a Millennial-friendly website and campaign focused on emerging artist talent: Zune Arts MySpace has added the feature to label music players on personal profiles to Zune-themed or a red Zune 8. Sales During its launch week, the original Zune, now Zune 30, was the second-most-sold portable media device with a 9% unit share in the United States: behind the market-leading iPod's 63%. For the first 6 months after launch, NPD Group figures show that the Zune 30 achieved approximately 10% market share in the Hard Drive-based MP3 market and 3% in the overall MP3 player market. According to Bloomberg Television 1.2 million Zune 30 players were sold between November 2006 and June 2007, surpassing a milestone. A price drop on Amazon.com during November 2007 temporarily boosted the brown Zune 30 to the top Sales Rank in electronics. On May 6, 2008, Microsoft announced that it had sold just over 2 million Zunes. Roughly one million of those were sold since the second generation Zunes launched in November 2007. On May 22, 2008, it was reported that GameStop "has decided to stop selling Microsoft's Zune players at its stores due to what it sees as insufficient demand from customers." A statement issued by Adam Sohn, Zune marketing manager said "We have a set of great partnerships...Best Buy, Target, Wal-Mart, and others." In January 2009, Microsoft's quarterly earnings filing with the SEC indicated that Zune sales had fallen $100 million from 2007 to 2008 during the fourth quarter of the calendar year. The Wall Street Journal estimated that sales appear to have dropped from about $185 million during the holiday period in 2007 to just $85 million in 2008. This may be due to the company's decision not to substantially update the Zune hardware in the fall of 2008. Zune market share decreased to 2% in the first half of 2009, according to the NPD Group. Availability outside the U.S. Microsoft released the Zune to Canadian consumers on June 13, 2008, marking the first time it was available outside the U.S. Microsoft has even made efforts to ban visitors outside the United States from Zune Originals. Users wishing to sign up for a Zune Tag could easily circumvent most problems by signing up for a US-based account. The Zune 2.0–3.* firmware does not support non-romanized fonts other than Cyrillic. East Asian characters used in Chinese and Japanese, for example, show up on the Zune device as mojibake instead of characters. Users have improvised ways to downgrade the firmware on the Zune device to older version that support Asian characters (V1 Zunes can be hacked to display Asian font). This is, however, not recommended by anyone. Legacy In 2012, Slate ran a reader poll inviting them to select a piece of defunct technology worthy of a reappraisal, where the Zune beat out Myspace, the PalmPilot, feature phones ("dumbphones") and older versions of Internet Explorer. Farhad Manjoo, the site's technology columnist, went to considerable lengths to acquire a Zune HD. "[It] wasn't as complete a failure as you may believe," he wrote. "If you purchased one over the iPod Touch back in 2009, you wouldn't have regretted it." Had Microsoft brought the first Zune to market at the same time as the iPod, or even the Zune HD at the time of the iPod Touch, it might have been a serious competitor to those players. Manjoo wrote: "By that point, iPod had become the world's de facto digital entertainment device," he recalled. "To beat it, Microsoft needed to offer something that would make Apple's device look pitifully old-fashioned. The Zune HD didn't do that. Its design marked it as being different from an iPod, but that was pretty much the only difference. There was no reason to buy the Zune unless you wanted to stand apart from the Apple cult. And there was a cost to standing apart from Apple: Because of its popularity, there were millions of apps and accessories for the iPod. As good as it was, the Zune HD couldn’t match Apple’s sheer market power." Nevertheless, it was an effective enough product that Manjoo had started using it as his backup music player ("when my iPhone’s battery dies"). Like the Edsel, to which it has sometimes been likened, Manjoo said the Zune's failure may have led the way to eventual success with other products. The HD's user interface, he noted, was the first such Microsoft product to rely on text rather than icons, and it would form the basis for Windows Phone, Windows 8, Xbox and all of the company's web-based services. The Segoe typeface that is now used in all of them, as well as Microsoft's current logo, was first used on the Zune. Business consultant Simon Sinek discusses the Zune in his 2019 book The Infinite Game, describing the Zune as a triumph of design but a failure of long-term strategic thinking. Microsoft was focused on beating Apple's portable music player, while Apple was behind the scenes focused on the iPhone that was introduced a year after the Zune's debut and, to a large extent, rendered obsolete handheld MP3 players by popularizing smart phones. At the end of the 2017 film Guardians of the Galaxy Vol. 2, Peter Quill is gifted a Zune by Kraglin to replace his destroyed Walkman. References External links Zune.net — Redirects to a Q&A for Zune users titled Important changes are coming to Zune Marketplace Zune Insider – Insider blog & podcast — Dead link Zune Podcast Support – An interview with Rob Greenlee, Podcast Programming Lead for the Zune about the second generation Zunes' support for podcasts. — Dead link Zune Podcast Connected Show Interview – An interview with Rob Greenlee. During the Interview we discuss his long career in Podcasting, his work in the Zune Podcast team, how to make podcasts successful and finally the ZuneHD Audiovisual introductions in 2006 Products and services discontinued in 2012 Consumer electronics brands Defunct consumer brands Discontinued Microsoft products Defunct online music stores Portable media players Digital audio players Windows Phone software
9593665
https://en.wikipedia.org/wiki/Liverpool%20Software%20Gazette
Liverpool Software Gazette
Liverpool Software Gazette was a short-lived computer magazine published by Microdigital Ltd, a company who were based in Liverpool, England and run by Bruce Everiss. History The magazine was in print for only eight issues of which the last was a double issue. Issues were bi-monthly from November 1979 to February 1981 though the last was actually dated February/April 1981. Initially costing 50p, the price increased to 75p by the fifth edition while the final double edition cost £1.50. The page count started at around 50 though by the fifth edition had reached 100 pages. The pressure of running both Microdigital and the magazine soon took its toll on the company, and the magazine was put up for sale during the final edition. It was sold to and incorporated into an Apple magazine where all non-Apple content was immediately dropped. The magazine was intended for an audience of sophisticated and experienced computer users and tackled a wide range of subjects from languages, machine code and CPUs, systems (both large and small), games, programming techniques, astronomy. In many cases the articles went far deeper than those normally tackled by the computer magazines of the day. Content There were some regular columns such as Pets Corner (for the Commodore PET), Apple Pips (for the Apple II), Nascom Notes and Nybbles (small BASIC tips and routines). Below are some of the contents from each of the issues. Note the general term for computers back then was Microcomputers (sometimes spelt as two words). In some cases the original spelling/typos have been left intact. Issue 1: Sargon meets the Nascom Programming Practices and Technics M5 System - an Interpreter for the Nascom One I'm Pilot, fly me Acorn Mastermind Pascal bytes the Apple Issue 2: Dungeons & Dragons Revisited Numerical Accuracy of Microcomputers Cesil - an introduction Acorn and the Kim Z-80 Processor Profile Revas & Zeap Application Software for Microcomputers Trekking by 'JTK' Byting more off your Disk Issue 3: AIM 65 Assembler Graphics Shapes (Series) Pilot Takes Off Pascal - an Introduction (series) Algol 68C on the Z80 Microcomputers and Biochemistry Sharp Machine Language Super Sort Issue 4: Star Gate Jet Set 6800 Processor Profile A Forth Introduction A Useful Pascal Program (series) A Marvel Called the MC6809 A Number Processor for the Acorn Architectural Software on the Cheap Commercial Micro Software Fundamentals Social Effects of Micro Computers Issue 5: Xtal Basic - The Extendable One Analysis of Systems Analysis Cesil Interpreted in Basic Algol 68 Fortran 77 Lisp Forms Processing Compiling Systems The Users View of Visicalc Issue 6 (Pascal special): 'Warning' Prolonged use of Pascal may seriously damage your mental health Integer Pascal on the Nascom Structured and not so Structured Programming TCL Pascal A Readers Contribution Alarming Your Computer The Romplus and Keyboard Filter Stargate Unlocked Tangerine Article MPL Language Issue 7 (CP/M special): Microcomputer Disk Operating Systems CP/M on the Sharp MZ-80K Z0 Article Graphics Software for the Apple II Introduction to Hi Resolution Graphics on the Apple Sharp PC 1211 Getting More From Your Genie Micro Chips in Use Now A Simple Machine Issue 8/9: Curing 'Clear' Keys Systems Programming with High Level Languages Macro's and Micro's Pilot Takes Off A Useful Romplus Programme M.P.L. Language Connecting the General Instruments AY-3-8910 Programmable Sound Generator to the 6502/6800 bus Sharp Basics Flashy Graphics Writing Decent Basic Programmes Microcomputer Communications for the Hobbyist Introduction to the Main Features of Algol 68 The TRS-80's Hidden Keyboard Apple D.O.S. 3.3 Notable articles The article in issue number 4, "Architectural software on the cheap" was contributed by Paul Coates, then working in the School of Architecture at Liverpool Polytechnic. In the opinion of the author, this was one of the earliest examples of the serious use of cheap micros for CAD. In those days the conventional wisdom was that CAD required a much larger computer and expensive graphics hardware. The article illustrated the use of the relatively cheap combination of a PET micro and an A3 plotter for generating floor plans, shadow casting and daylight calculation. The work illustrated was funded by a small RIBA (The Royal Institute of British Architects) grant for teaching architecture using computers. The article on TCL Pascal in issue 6 was by Anne Scott and John Stout. External links Article Architectural software on the cheap from Liverpool Software Gazette issue 4 Computer magazine history featuring Liverpool Software Gazette Profile of Bruce Everiss PDFs of four issues can be found here Microdigital in Liverpool from Bruceongames Article by Bruce Everiss giving background to Liverpool Software Gazette Home computer magazines Defunct computer magazines published in the United Kingdom Magazines established in 1979 Magazines disestablished in 1981 Bi-monthly magazines published in the United Kingdom Mass media in Liverpool
13940627
https://en.wikipedia.org/wiki/Black%20Zero
Black Zero
Black Zero is a name shared by two supervillains, two terrorist organizations, one special forces group, and a computer virus that have all appeared in various comic book series published by DC Comics. Original Black Zero supervillain Fictional character biography The original Black Zero appeared Pre-Crisis in Superman #205 (1968), in a story entitled "The Man Who Destroyed Krypton" written by Otto Binder with artwork by Al Plastino. This Black Zero revealed to Superman that he was a space saboteur who destroyed planets. He had been hired to destroy Krypton, and discovered that the internal stresses which would have destroyed it in any case were dying down, necessitating his involvement to ensure it exploded. In the present day, Black Zero came to Earth, threatening to destroy it as he did Krypton. In desperation, Superman released Jax-Ur, a prisoner of the Phantom Zone, who wished to avenge Krypton's destruction. As he launched a devastatingly powerful missile toward Earth, Black Zero attacked Jax-Ur with a red kryptonite bullet, causing his body to mutate wildly into several serpent-like forms. This proved to be Black Zero's undoing; while Superman saved the Earth from the missile, Jax-Ur transformed into a Medusa-like form, turning Black Zero to stone with his gaze. Black Zero's body was then shattered by Jax-Ur, in recompense for what he had done to Krypton. This story suggests that, were it not for something none of them knew, Jor-El would have been wrong and the Science Council right. E. Nelson Bridwell hypothesized that what Black Zero had noticed was the Green Lantern Tomar Re's attempts to prevent the destruction (as seen in Superman #257 (1972)), meaning the two sets of interference canceled each other out, and Jor-El was right after all. Power and abilities Black Zero had the ability to create "psycho-molecules" with his mind, which he could then shape into any form of matter he chose. His brain had been fitted with a plastic coating, preventing him from being hypnotised. He could make his body intangible, enabling him to pass through solid objects. He was a skilled engineer, able to create weapons and devices that could destroy entire planets. He was also skilled in the art of disguise. Black Zero organizations The first Post-Crisis version of Black Zero appeared in the 1988 World of Krypton miniseries written by John Byrne and illustrated by Mike Mignola. Although later described as a "clone liberation movement" in Superboy (vol. 4) #61, Black Zero was described as a "terrorist" organization that was ultimately responsible for the destruction of Krypton. During the Third Age of Krypton, Kryptonians extended their lives by maintaining clones in suspended animation (the Clone Banks), which they then harvested for body parts. Trouble in Kryptonian society concerning this issue emerged after a prominent citizen named Nyra removed one of her clones from stasis in order to marry the clone to her own son. The enraged son after killing his clone fiancé, publicly confronted his mother and killed her as well, but was forcibly stopped from killing himself. After widespread public knowledge ensued concerning the fact that a clone had been able to achieve full sentience as a living being, a new Kryptonian war began. Prominently featured in this war was the Black Zero organization, which acted against this "genetic slavery"; this began the War of Clone Rights, which lasted for a thousand years. In their final act, Black Zero detonated a device that would later become known as the Destroyer. In essence, this device functioned as a giant nuclear gun, that fired a concentrated, sustained burst of nuclear energy directly into Krypton's core. Though subsequently destroyed by Van-L, an ancestor of Jor-El himself, the Destroyer's effect would later be fully realized; it causes a chain reaction deep within Krypton that would, thousands of years later, obliterate the planet. A second Black Zero group appeared in the 2005 graphic novel Superman: Infinite City. Most recently, the name "Black Zero" was used by an elite Kryptonian military force, under the command of Ursa. This unit included Thara Ak-Var's parents. The unit, except for Ursa, was killed attempting to defend Kandor from Brainiac. The "Black Zero II" computer virus The Black Zero virus appeared in Superman Plus/Legion of Super-Heroes #1, a 1997 one-shot teaming Superman with the Legion of Super-Heroes. An intelligent computer virus created by the terrorist group, Black Zero II was dormant in the Kryptonian technology used to create the Fortress of Solitude. After being inadvertently activated by Apparition, Black Zero II ran Superman and the Legion through a warped version of Kryptonian history, intending to finish with the destruction of Earth. It was halted when Brainiac 5 inadvertently caused a power outage, and subsequently removed the virus from the Fortress' computers. Black Zero (Superboy) The second Black Zero is a fictional character, a supervillain from the DC Comics universe. The character first appeared in Superboy Vol. 4 #61 as part of the '"Hypertension" storyline and was created by Karl Kesel and Tom Grummett. Fictional character biography Black Zero was an alternate version of Superboy from a reality in which Superman never returned from the dead after his battle with Doomsday. This reality mirrored the main continuity with Paul Westfield, the Executive Director of Project Cadmus, secretly using his own cellstock for his team of scientists to immediately create a human clone that would be genetically engineered to look like Superman with powers deriving from a telekinetic field. Instead of the Newsboy Legion clones releasing him prematurely, the clone went through the full artificial maturation process to match the age of the original and became the new Superman. However, he believed that he was perceived as a usurper and resented by the people for not being the original Superman. After a battle with his reality's versions of Brainiac, Maxima, and Metallo led to the deaths of Supergirl, Brainiac and 318 civilians, a public backlash began against the new Superman and other clones that the people dubbed "genetix"; this led to Congress demanding the shutdown of Project Cadmus with the Guardian and the original directors of Project Cadmus sans Westfield abandoning the project. After the Guardian was killed by an anti-genetix mob, the new Superman retreated to the original Superman's Fortress of Solitude and sought guidance from Kryptonian history. When he learned of the Black Zero clone liberation movement from Krypton, he discarded the name of Superman and took Black Zero as his new name. He decided to defend genetix everywhere and stop the anti-clone lobby by any means necessary. Turning Project Cadmus into Fortress Cadmus and creating an army of Guardian clones known as Stormguards with no previous memories of the original Guardian, Black Zero began a war in which most of his Earth's heroes were destroyed and cloned as soldiers for Cadmus with extreme changes by his head geneticist Dabney Donovan. After conquering his Earth, Black Zero became aware of the existence of other realities when he encountered the Challengers of the Unknown, who arrived to his homeworld from a dimensional doorway. Metron of the New Gods, who appeared shortly after the Challengers, struck an agreement with Black Zero and offered him the chance to enter Hypertime via the construction of a reactor powered by the element Hyperium. Black Zero "rescued" other worlds in which he felt that genetix were being oppressed and left the conquered world's version of Paul Westfield in charge of the genetix on that world. The Challengers attempted to stop him by launching commando raids against Black Zero and his forces and smuggling themselves to whatever world was next in his crusade. To Black Zero's bafflement, the Superboys of these worlds often stood against him, leading him to store them in stasis chambers similar to the clone banks of Krypton. A Superboy from another reality that Black Zero tried to conquer escaped with a stolen Hyperjacket from Fortress Cadmus, but was shot with a laser by the Paul Westfield of his reality resulting in fatal burns. He warned the Justice League of the main continuity of Black Zero's coming before he died. Learning that the deceased Superboy was a genetic match and using the Hyperjacket repaired by the Gadget Guru of the Hairies, the Superboy of the main continuity traveled to other worlds where he learned of Black Zero and eventually landed on an Earth in which Superboy was Kal-El. After Superboy was in possession of the hyperjacket, Black Zero followed him to each world that he visited, capturing all of the versions of Superboy that Kon-El encountered. When he reached the world of the Kal-El Superboy, he captured Krypto and ambushed both Kon-El and Kal-El using his Stormguard forces and his own powers. Learning that Superboy wasn't the Superboy that stole the hyperjacket, Black Zero sent him to Fortress Cadmus and revealed his true identity and told his story to Superboy via the latter in a mnemonic chair that showed parts of Black Zero's life in virtual reality. He revealed his plans to take over Superboy's reality, which he used a supposed anti-clone sentiment as justification, and to use one of the Doomsdays that he collected from alternate realities against Superman. After revealing his plan, a reformed version of Knockout from the deceased Superboy's world attacked with the Challengers of The Unknown, posing as Stormguards, and went after her reality's Paul Westfield. Westfield leaped into the Hyperium reactor and unknowingly erased every Paul Westfield in existence as well as himself. Angry at the loss of Westfield, who he considered a father figure, Black Zero sent Superboy, Knockout, and the Challengers to the Doomsday chamber to be destroyed while he prepared to take over the next world. Superboy and the Challengers escaped with Knockout making the ultimate sacrifice against the Doomsdays. Black Zero was eventually defeated by a combined effort from Superboy, the Challengers of the Unknown, and all the other Superboys imprisoned at Fortress Cadmus. After Metron sends the Superboys and Stormguards back to their own realities along with immobilizing Black Zero with a Mother Box, he provides a ship for the Challengers and Superboy to fly through the hyperium reactor and into a hyperstorm to return home. Superboy intended to take Black Zero to his homeworld to keep watch over him, but Black Zero attacks the ship with his tactile telekinesis. Superboy dives into the hyperstorm and is almost struck by it until Black Zero rescued Superboy and saves the Challengers except for Matthew "Red" Ryan. Superboy attempted to convince Black Zero that his mission was wrong and to give people a chance, but Black Zero is engulfed by hyperstorm energy and disappears before responding to Superboy. After his "death", Black Zero became a reminder to Superboy as to what could happen if he ever went out of control. During a battle with Etrigan over an enchanted armored gauntlet known as "Hyssa's Fist", Superboy puts on the gauntlet and while his strength increased, his costume was transformed into an exact copy of Black Zero's costume with the S-shield on the gold harness. He quickly used his tactile telekinesis to disassemble the gauntlet from his right hand. Powers, abilities, and equipment Tactile telekinesis Black Zero's main unique power is an advanced form of "tactile telekinesis". Like Superboy, it is a telekinetic force field that surrounds his body and allows him to simulate superhuman speed, strength, flight and invulnerability. Tactile telekinesis also allows him to disassemble machines and other complex constructions by touch. Unlike Superboy, Black Zero's field provides him with resistance to energy attacks and he could use tactile telekinesis to telekinetically hold an opponent immobile. Powers not derived from tactile telekinesis As he explained to Superboy, Black Zero's human DNA was made to be genetically similar to that of Superman's DNA. Like Superboy, he is a living solar battery with his tactile telekinetic powers being fueled by solar energy. Grown to a full adult by the Project Cadmus of his reality, Black Zero gained powers similar to Superman due to his advanced age such as heat vision and super hearing. With the combination of his tactile telekinesis and his simulated Kryptonian powers it is implied that Black Zero is more powerful than Superman. He easily seems to hold his own against multiple versions of Superboys, even the vastly powered Pre-Crisis Earth-1 version. Black Zero also seems to be well prepared to conquer a reality where he will have to fight Superman after his army of Doomsdays are disposed of. Costumes and equipment Black Zero wears a black costume with red accents offset by a red cape, red belt at the waist, gold harness with a black "0" on his chest, gold belt on the right leg, and a pair of black glasses. Additionally, he wears a "hypergun" on his right arm equipped with hyper-bullets which he can fire at his targets and instantly transport them via hypertravel directly to his stasis cells at Fortress Cadmus. In other media Television In the Smallville episode "Kandor", Major Zod and his armies went to war with the terrorist group Black Zero, who later destroyed Kandor. The Black Zero terrorist organization is a significant aspect of the plot of the TV series Krypton, led by Jax-Ur aka Sela-Sonn after she was expelled from the Science Guild. Film Black Zero is heavily featured in Man of Steel. In this version, The Black Zero was a Kryptonian prison frigate originally designed by Jor-El to transport criminals into the Phantom Zone. It was used to imprison General Zod and his battalion the Sword of Rao following their failed coup, and, following Krypton's destruction, was later repurposed as their command ship. General Zod proceeded to utilize the ship as a means of searching the universe for Kryptonian survivors. When Kal-El triggered the distress beacon in his discovered Scout Ship 0344, the Sword of Rao travelled to Earth in their resumed pursuit of the Growth Codex. Upon finally reaching Earth, Zod attempted to use a World Engine in the Indian Ocean to terraform the planet into a world suitable for Kryptonians, but was defeated by Emil Hamilton's activation of the Phantom Drive aboard Kal-El's Starcraft, sucking the ship and Zod's soldiers back into the Phantom Zone above Metropolis. Black Zero is seen again in the film's flashback Batman v Superman: Dawn of Justice, when Bruce Wayne arrives in his helicopter upon learning of the Metropolis invasion, heading to the Wayne Financial Building before it collapses, killing the dozens of people left inside, much to the horror of Bruce Wayne, who saw it happen. See also Alternative versions of Superman Comic book death Project Cadmus Superboy (Kon-El) Superman References External links Supermanica entry on the Pre-Crisis Black Zero Characters created by Karl Kesel DC Comics supervillains DC Comics characters with superhuman strength Fictional genetically engineered characters Fictional clones Fictional mass murderers Kryptonians Comic book terrorist organizations Comics characters introduced in 1968 Comics characters introduced in 1999 Space pirates Characters created by Otto Binder Characters created by John Byrne (comics) Characters created by Mike Mignola Superman characters de:Superboy#Black Zero
730243
https://en.wikipedia.org/wiki/Social%20engineering%20%28security%29
Social engineering (security)
In the context of information security, social engineering is the psychological manipulation of people into performing actions or divulging confidential information. This differs from social engineering within the social sciences, which does not concern the divulging of confidential information. A type of confidence trick for the purpose of information gathering, fraud, or system access, it differs from a traditional "con" in that it is often one of many steps in a more complex fraud scheme. It has also been defined as "any act that influences a person to take an action that may or may not be in their best interests." An example of social engineering is the use of the "forgot password" function on most websites which require login. An improperly-secured password-recovery system can be used to grant a malicious attacker full access to a user's account, while the original user will lose access to the account. Information security culture Employee behavior can have a big impact on information security in organizations. Cultural concepts can help different segments of the organization work effectively or work against effectiveness towards information security within an organization. "Exploring the Relationship between Organizational Culture and Information Security Culture" provides the following definition of information security culture: "ISC is the totality of patterns of behavior in an organization that contribute to the protection of information of all kinds." Andersson and Reimers (2014) found that employees often do not see themselves as part of the organization Information Security "effort" and often take actions that ignore organizational information security best interests. Research shows Information security culture needs to be improved continuously. In "Information Security Culture from Analysis to Change," authors commented that "it's a never ending process, a cycle of evaluation and change or maintenance." They suggest that to manage information security culture, five steps should be taken: Pre-evaluation, strategic planning, operative planning, implementation, and post-evaluation. Pre-Evaluation: to identify the awareness of information security within employees and to analyse current security policy. Strategic Planning: to come up with a better awareness-program, we need to set clear targets. Clustering people is helpful to achieve it. Operative Planning: set a good security culture based on internal communication, management-buy-in, and security awareness and training program. Implementation: four stages should be used to implement the information security culture. They are commitment of the management, communication with organizational members, courses for all organizational members, and commitment of the employees. Techniques and terms All social engineering techniques are based on specific attributes of human decision-making known as cognitive biases. These biases, sometimes called "bugs in the human hardware,” are exploited in various combinations to create attack techniques, some of which are listed below. The attacks used in social engineering can be used to steal employees' confidential information. The most common type of social engineering happens over the phone. Other examples of social engineering attacks are criminals posing as exterminators, fire marshals and technicians to go unnoticed as they steal company secrets. One example of social engineering is an individual who walks into a building and posts an official-looking announcement to the company bulletin that says the number for the help desk has changed. So, when employees call for help the individual asks them for their passwords and IDs thereby gaining the ability to access the company's private information. Another example of social engineering would be that the hacker contacts the target on a social networking site and starts a conversation with the target. Gradually the hacker gains the trust of the target and then uses that trust to get access to sensitive information like password or bank account details. Social engineering relies heavily on the six principles of influence established by Robert Cialdini. Cialdini's theory of influence is based on six key principles: reciprocity, commitment and consistency, social proof, authority, liking, scarcity. Six key principles Authority In social engineering, the attacker may pose as authority to increase the likelihood of adherence from the victim. Intimidation Attacker (potentially disguised) informs or implies that there will be negative consequences if certain actions are not performed. Consequences could include subtle intimidation phrases such as "I'll tell your manager" to much worse. Consensus/Social proof People will do things that they see other people are doing. For example, in one experiment, one or more confederates would look up into the sky; bystanders would then look up into the sky to see what they were missing. At one point this experiment was aborted, as so many people were looking up that they stopped traffic. See conformity, and the Asch conformity experiments. Scarcity Perceived scarcity will generate demand. The common advertising phrase "while supplies last" capitalizes on a sense of scarcity. Urgency Linked to scarcity, attackers use urgency as a time-based psychological principle of social engineering. For example, saying offers are available for a "limited time only" encourages sales through a sense of urgency. Familiarity / Liking People are easily persuaded by other people whom they like. Cialdini cites the marketing of Tupperware in what might now be called viral marketing. People were more likely to buy if they liked the person selling it to them. Some of the many biases favoring more attractive people are discussed. See physical attractiveness stereotype. Four social engineering vectors Vishing Vishing, otherwise known as "voice phishing", is the criminal practice of using social engineering over a telephone system to gain access to private personal and financial information from the public for the purpose of financial reward. It is also employed by attackers for reconnaissance purposes to gather more detailed intelligence on a target organization. Phishing Phishing is a technique of fraudulently obtaining private information. Typically, the phisher sends an e-mail that appears to come from a legitimate business—a bank, or credit card company—requesting "verification" of information and warning of some dire consequence if it is not provided. The e-mail usually contains a link to a fraudulent web page that seems legitimate—with company logos and content—and has a form requesting everything from a home address to an ATM card's PIN or a credit card number. For example, in 2003, there was a phishing scam in which users received emails supposedly from eBay claiming that the user's account was about to be suspended unless a link provided was clicked to update a credit card (information that the genuine eBay already had). By mimicking a legitimate organization's HTML code and logos, it is relatively simple to make a fake Website look authentic. The scam tricked some people into thinking that eBay was requiring them to update their account information by clicking on the link provided. By indiscriminately spamming extremely large groups of people, the "phisher" counted on gaining sensitive financial information from the small percentage (yet large number) of recipients who already have eBay accounts and also fall prey to the scam. Smishing The act of using SMS text messaging to lure victims into a specific course of action. Like phishing it can be clicking on a malicious link or divulging information. Examples are text messages that claim to be from a common carrier (like FedEx) stating a package is in transit, with a link provided. Impersonation Pretending or pretexting to be another person with the goal of gaining access physically to a system or building. Impersonation is used in the "SIM swap scam" fraud. Other concepts Pretexting Pretexting (adj. pretextual) is the act of creating and using an invented scenario (the pretext) to engage a targeted victim in a manner that increases the chance the victim will divulge information or perform actions that would be unlikely in ordinary circumstances. An elaborate lie, it most often involves some prior research or setup and the use of this information for impersonation (e.g., date of birth, Social Security number, last bill amount) to establish legitimacy in the mind of the target. As a background, pretexting can be interpreted as the first evolution of social engineering, and continued to develop as social engineering incorporated current-day technologies. Current and past examples of pretexting demonstrate this development. This technique can be used to fool a business into disclosing customer information as well as by private investigators to obtain telephone records, utility records, banking records and other information directly from company service representatives. The information can then be used to establish even greater legitimacy under tougher questioning with a manager, e.g., to make account changes, get specific balances, etc. Pretexting can also be used to impersonate co-workers, police, bank, tax authorities, clergy, insurance investigators—or any other individual who could have perceived authority or right-to-know in the mind of the targeted victim. The pretexter must simply prepare answers to questions that might be asked by the victim. In some cases, all that is needed is a voice that sounds authoritative, an earnest tone, and an ability to think on one's feet to create a pretextual scenario. Vishing Phone phishing (or "vishing") uses a rogue interactive voice response (IVR) system to recreate a legitimate-sounding copy of a bank or other institution's IVR system. The victim is prompted (typically via a phishing e-mail) to call in to the "bank" via a (ideally toll free) number provided in order to "verify" information. A typical "vishing" system will reject log-ins continually, ensuring the victim enters PINs or passwords multiple times, often disclosing several different passwords. More advanced systems transfer the victim to the attacker/defrauder, who poses as a customer service agent or security expert for further questioning of the victim. Spear phishing Although similar to "phishing", spear phishing is a technique that fraudulently obtains private information by sending highly customized emails to few end users. It is the main difference between phishing attacks because phishing campaigns focus on sending out high volumes of generalized emails with the expectation that only a few people will respond. On the other hand, spear-phishing emails require the attacker to perform additional research on their targets in order to "trick" end users into performing requested activities. The success rate of spear-phishing attacks is considerably higher than phishing attacks with people opening roughly 3% of phishing emails when compared to roughly 70% of potential attempts. When users actually open the emails phishing emails have a relatively modest 5% success rate to have the link or attachment clicked when compared to a spear-phishing attack's 50% success rate. Spear-phishing success is heavily dependent on the amount and quality of OSINT (open-source intelligence) that the attacker can obtain. Social media account activity is one example of a source of OSINT. Water holing Water holing is a targeted social engineering strategy that capitalizes on the trust users have in websites they regularly visit. The victim feels safe to do things they would not do in a different situation. A wary person might, for example, purposefully avoid clicking a link in an unsolicited email, but the same person would not hesitate to follow a link on a website they often visit. So, the attacker prepares a trap for the unwary prey at a favored watering hole. This strategy has been successfully used to gain access to some (supposedly) very secure systems. The attacker may set out by identifying a group or individuals to target. The preparation involves gathering information about websites the targets often visit from the secure system. The information gathering confirms that the targets visit the websites and that the system allows such visits. The attacker then tests these websites for vulnerabilities to inject code that may infect a visitor's system with malware. The injected code trap and malware may be tailored to the specific target group and the specific systems they use. In time, one or more members of the target group will get infected and the attacker can gain access to the secure system. Baiting Baiting is like the real-world Trojan horse that uses physical media and relies on the curiosity or greed of the victim. In this attack, attackers leave malware-infected floppy disks, CD-ROMs, or USB flash drives in locations people will find them (bathrooms, elevators, sidewalks, parking lots, etc.), give them legitimate and curiosity-piquing labels, and wait for victims. For example, an attacker may create a disk featuring a corporate logo, available from the target's website, and label it "Executive Salary Summary Q2 2012". The attacker then leaves the disk on the floor of an elevator or somewhere in the lobby of the target company. An unknowing employee may find it and insert the disk into a computer to satisfy their curiosity, or a good Samaritan may find it and return it to the company. In any case, just inserting the disk into a computer installs malware, giving attackers access to the victim's PC and, perhaps, the target company's internal computer network. Unless computer controls block infections, insertion compromises PCs "auto-running" media. Hostile devices can also be used. For instance, a "lucky winner" is sent a free digital audio player compromising any computer it is plugged to. A "road apple" (the colloquial term for horse manure, suggesting the device's undesirable nature) is any removable media with malicious software left in opportunistic or conspicuous places. It may be a CD, DVD, or USB flash drive, among other media. Curious people take it and plug it into a computer, infecting the host and any attached networks. Again, hackers may give them enticing labels, such as "Employee Salaries" or "Confidential". One study done in 2016 had researchers drop 297 USB drives around the campus of the University of Illinois. The drives contained files on them that linked to webpages owned by the researchers. The researchers were able to see how many of the drives had files on them opened, but not how many were inserted into a computer without having a file opened. Of the 297 drives that were dropped, 290 (98%) of them were picked up and 135 (45%) of them "called home". Quid pro quo Quid pro quo means something for something: An attacker calls random numbers at a company, claiming to be calling back from technical support. Eventually this person will hit someone with a legitimate problem, grateful that someone is calling back to help them. The attacker will "help" solve the problem and, in the process, have the user type commands that give the attacker access or launch malware. In a 2003 information security survey, 91% of office workers gave researchers what they claimed was their password in answer to a survey question in exchange for a cheap pen. Similar surveys in later years obtained similar results using chocolates and other cheap lures, although they made no attempt to validate the passwords. Tailgating An attacker, seeking entry to a restricted area secured by unattended, electronic access control, e.g. by RFID card, simply walks in behind a person who has legitimate access. Following common courtesy, the legitimate person will usually hold the door open for the attacker or the attackers themselves may ask the employee to hold it open for them. The legitimate person may fail to ask for identification for any of several reasons, or may accept an assertion that the attacker has forgotten or lost the appropriate identity token. The attacker may also fake the action of presenting an identity token. Other types Common confidence tricksters or fraudsters also could be considered "social engineers" in the wider sense, in that they deliberately deceive and manipulate people, exploiting human weaknesses to obtain personal benefit. They may, for example, use social engineering techniques as part of an IT fraud. As of the early 2000s, another type of social engineering technique includes spoofing or hacking IDs of people having popular e-mail IDs such as Yahoo!, Gmail, or Hotmail. Additionally, some spoofing attempts included emails from major online service providers, like PayPal. This led to the "proposed standard" of Sender Policy Framework RFC 7208 dated April 2014, in combination with DMARC, as means to combat spoofing. Among the many motivations for this deception are: Phishing credit-card account numbers and their passwords. Cracking private e-mails and chat histories, and manipulating them by using common editing techniques before using them to extort money and creating distrust among individuals. Cracking websites of companies or organizations and destroying their reputation. Computer virus hoaxes Convincing users to run malicious code within the web browser via self-XSS attack to allow access to their web account Another type is to read sensitive information of unshielded or unprotected Displays and input devices, called Shoulder surfing. Countermeasures Organizations reduce their security risks by: Training to Employees: Training employees in security protocols relevant to their position. (e.g., in situations such as tailgating, if a person's identity cannot be verified, then employees must be trained to politely refuse.) Standard Framework: Establishing frameworks of trust on an employee/personnel level (i.e., specify and train personnel when/where/why/how sensitive information should be handled) Scrutinizing Information: Identifying which information is sensitive and evaluating its exposure to social engineering and breakdowns in security systems (building, computer system, etc.) Security Protocols: Establishing security protocols, policies, and procedures for handling sensitive information. Event Test: Performing unannounced, periodic tests of the security framework. Inoculation: Preventing social engineering and other fraudulent tricks or traps by instilling a resistance to persuasion attempts through exposure to similar or related attempts. Review: Reviewing the above steps regularly: no solutions to information integrity are perfect. Waste Management: Using a waste management service that has dumpsters with locks on them, with keys to them limited only to the waste management company and the cleaning staff. Locating the dumpster either in view of employees so that trying to access it carries a risk of being seen or caught, or behind a locked gate or fence where the person must trespass before they can attempt to access the dumpster. The lifecycle of social engineering Information gathering: Information gathering is the first and foremost step of the lifecycle. It requires much patience and keenly watching habits of the victim. This step gathering data about the victim's interests, personal information. It determines the success rate of the overall attack. Engaging with victim: After gathering required amount of information, the attacker opens a conversation with the victim smoothly without the victim finding anything inappropriate. Attacking: This step generally occurs after a long period of engaging with the target and during this information from the target is retrieved by using social engineering. In phase, the attacker gets the results from the target. Closing interaction: This is the last step which includes slowly shutting down the communication by the attacker without arising any suspicion in the victim. In this way, the motive is fulfilled as well as the victim rarely comes to know the attack even happened. Notable social engineers Frank Abagnale Jr. Frank Abagnale Jr. is an American security consultant known for his background as a former con man, check forger, and impostor while he was between the ages of 15 and 21. He became one of the most notorious impostors, claiming to have assumed no fewer than eight identities, including an airline pilot, a physician, a U.S. Bureau of Prisons agent, and a lawyer. Abagnale escaped from police custody twice (once from a taxiing airliner and once from a U.S. federal penitentiary) before turning 22 years old. The popular Steven Spielberg movie Catch Me If You Can is based on his life. Kevin Mitnick Kevin Mitnick is an American computer security consultant, author and hacker, best known for his high-profile 1995 arrest and later five-year conviction for various computer and communications-related crimes. Susan Headley Susan Headley was an American hacker active during the late 1970s and early 1980s widely respected for her expertise in social engineering, pretexting, and psychological subversion. She was known for her specialty in breaking into military computer systems, which often involved going to bed with military personnel and going through their clothes for usernames and passwords while they slept. She became heavily involved in phreaking with Kevin Mitnick and Lewis de Payne in Los Angeles, but later framed them for erasing the system files at US Leasing after a falling out, leading to Mitnick's first conviction. She retired to professional poker. James Linton James Linton is a British hacker and social engineer who in 2017 used OSINT and spear phishing techniques to trick a variety of targets over email including the CEOs of Major Banks, and members of the Trump White House Administration. He then went to work in email security where he socially engineered BEC (Business Email Compromise) threat actors to collect specific threat intelligence. Badir Brothers Brothers Ramy, Muzher, and Shadde Badir—all of whom were blind from birth—managed to set up an extensive phone and computer fraud scheme in Israel in the 1990s using social engineering, voice impersonation, and Braille-display computers. Christopher J. Hadnagy Christopher J. Hadnagy is an American social engineer and information technology security consultant. He is best known as an author of 4 books on social engineering and cyber security and founder of Innocent Lives Foundation, an organization that helps tracking and identifying child trafficking using various security techniques such as seeking the assistance of information security specialists, utilizing data from open-source intelligence (OSINT) and collaborating with law enforcement. Law In common law, pretexting is an invasion of privacy tort of appropriation. Pretexting of telephone records In December 2006, United States Congress approved a Senate sponsored bill making the pretexting of telephone records a federal felony with fines of up to $250,000 and ten years in prison for individuals (or fines of up to $500,000 for companies). It was signed by President George W. Bush on 12 January 2007. Federal legislation The 1999 "GLBA" is a U.S. Federal law that specifically addresses pretexting of banking records as an illegal act punishable under federal statutes. When a business entity such as a private investigator, SIU insurance investigator, or an adjuster conducts any type of deception, it falls under the authority of the Federal Trade Commission (FTC). This federal agency has the obligation and authority to ensure that consumers are not subjected to any unfair or deceptive business practices. US Federal Trade Commission Act, Section 5 of the FTCA states, in part: "Whenever the Commission shall have reason to believe that any such person, partnership, or corporation has been or is using any unfair method of competition or unfair or deceptive act or practice in or affecting commerce, and if it shall appear to the Commission that a proceeding by it in respect thereof would be to the interest of the public, it shall issue and serve upon such person, partnership, or corporation a complaint stating its charges in that respect." The statute states that when someone obtains any personal, non-public information from a financial institution or the consumer, their action is subject to the statute. It relates to the consumer's relationship with the financial institution. For example, a pretexter using false pretenses either to get a consumer's address from the consumer's bank, or to get a consumer to disclose the name of their bank, would be covered. The determining principle is that pretexting only occurs when information is obtained through false pretenses. While the sale of cell telephone records has gained significant media attention, and telecommunications records are the focus of the two bills currently before the United States Senate, many other types of private records are being bought and sold in the public market. Alongside many advertisements for cell phone records, wireline records and the records associated with calling cards are advertised. As individuals shift to VoIP telephones, it is safe to assume that those records will be offered for sale as well. Currently, it is legal to sell telephone records, but illegal to obtain them. 1st Source Information Specialists U.S. Rep. Fred Upton (R-Kalamazoo, Michigan), chairman of the Energy and Commerce Subcommittee on Telecommunications and the Internet, expressed concern over the easy access to personal mobile phone records on the Internet during a House Energy & Commerce Committee hearing on "Phone Records For Sale: Why Aren't Phone Records Safe From Pretexting?" Illinois became the first state to sue an online records broker when Attorney General Lisa Madigan sued 1st Source Information Specialists, Inc. A spokeswoman for Madigan's office said. The Florida-based company operates several Web sites that sell mobile telephone records, according to a copy of the suit. The attorneys general of Florida and Missouri quickly followed Madigan's lead, filing suits respectively, against 1st Source Information Specialists and, in Missouri's case, one other records broker – First Data Solutions, Inc. Several wireless providers, including T-Mobile, Verizon, and Cingular filed earlier lawsuits against records brokers, with Cingular winning an injunction against First Data Solutions and 1st Source Information Specialists. U.S. Senator Charles Schumer (D-New York) introduced legislation in February 2006 aimed at curbing the practice. The Consumer Telephone Records Protection Act of 2006 would create felony criminal penalties for stealing and selling the records of mobile phone, landline, and Voice over Internet Protocol (VoIP) subscribers. HP Patricia Dunn, former chairwoman of Hewlett Packard, reported that the HP board hired a private investigation company to delve into who was responsible for leaks within the board. Dunn acknowledged that the company used the practice of pretexting to solicit the telephone records of board members and journalists. Chairman Dunn later apologized for this act and offered to step down from the board if it was desired by board members. Unlike Federal law, California law specifically forbids such pretexting. The four felony charges brought on Dunn were dismissed. Preventive measures Taking some precautions reduce the risk of being a victim to social engineering frauds. The precautions that can be made are as follows: Be aware of offers that seem "Too good to be true". Use multifactor authentication. Avoid clicking on attachments from unknown sources. Not giving out personal information to anyone via email, phone, or text messages. Use of spam filter software. Avoid befriending people that you do not know in real life. Teach kids to contact a trusted adult in case they are being bullied over the internet (cyberbullying) or feel threatened by anything online. See also (CSEPS) s, which often use similar tactics (though usually not for criminal purposes) Cyber security awareness References Further reading Boyington, Gregory. (1990). 'Baa Baa Black Sheep' Published by Gregory Boyington Harley, David. 1998 Re-Floating the Titanic: Dealing with Social Engineering Attacks EICAR Conference. Laribee, Lena. June 2006 Development of methodical social engineering taxonomy project Master's Thesis, Naval Postgraduate School. Leyden, John. 18 April 2003. Office workers give away passwords for a cheap pen. The Register. Retrieved 2004-09-09. Long, Johnny. (2008). No Tech Hacking – A Guide to Social Engineering, Dumpster Diving, and Shoulder Surfing Published by Syngress Publishing Inc. Mann, Ian. (2008). Hacking the Human: Social Engineering Techniques and Security Countermeasures Published by Gower Publishing Ltd. or Mitnick, Kevin, Kasperavičius, Alexis. (2004). CSEPS Course Workbook. Mitnick Security Publishing. Mitnick, Kevin, Simon, William L., Wozniak, Steve,. (2002). The Art of Deception: Controlling the Human Element of Security Published by Wiley. or Hadnagy, Christopher, (2011) Social Engineering: The Art of Human Hacking Published by Wiley. N.J. Evans. (2009). "Information Technology Social Engineering: An Academic Definition and Study of Social Engineering-Analyzing the Human Firewall." Graduate Theses and Dissertations. 10709. https://lib.dr.iastate.edu/etd/10709 Z. Wang, L. Sun and H. Zhu. (2020) "Defining Social Engineering in Cybersecurity," in IEEE Access, vol. 8, pp. 85094-85115, doi: 10.1109/ACCESS.2020.2992807. External links Social Engineering Fundamentals – Securityfocus.com. Retrieved 3 August 2009. Should Social Engineering be a part of Penetration Testing? – Darknet.org.uk. Retrieved 3 August 2009. "Protecting Consumers' Phone Records", Electronic Privacy Information Center US Committee on Commerce, Science, and Transportation . Retrieved 8 February 2006. Plotkin, Hal. Memo to the Press: Pretexting is Already Illegal. Retrieved 9 September 2006. Cybercrime
16476828
https://en.wikipedia.org/wiki/4543%20Phoinix
4543 Phoinix
4543 Phoinix is a large Jupiter trojan from the Greek camp, approximately in diameter. It was discovered on 2 February 1989, by American astronomer Carolyn Shoemaker at the Palomar Observatory in California. The assumed C-type asteroid is possibly elongated and has a longer-than-average rotation period of 38.9 hours. It is one of the 60 largest Jupiter trojans and was named after Phoenix (Phoinix) from Greek mythology. Orbit and classification Phoinix is a dark Jovian asteroid orbiting in the leading Greek camp at Jupiter's Lagrangian point, 60° ahead of the Gas Giant's orbit in a 1:1 resonance (see Trojans in astronomy). This asteroid is not a member of any asteroid family but belongs to the Jovian background population. It orbits the Sun at a distance of 4.6–5.6 AU once every 11 years and 7 months (4,237 days; semi-major axis of 5.12 AU). Its orbit has an eccentricity of 0.10 and an inclination of 15° with respect to the ecliptic. The asteroid was first observer as at Heidelberg Observatory in February 1930. The body's observation arc begins with a precovery taken at Palomar in April 1955, nearly 34 years prior to its official discovery observation. Physical characteristics Phoinix is a generically assumed C-type asteroid. Its V–I color index of 1.20 is the highest of all larger Jovian trojans, which are typically D-types with a V–I index near 0.90 (also see table below). Rotation period In November 2011, a rotational lightcurve of Phoinix was obtained from photometric observations by Italian astronomer Albino Carbognani at the OAVdA Observatory , Italy, with follow-up observations conducted at Calar Alto Observatory by Stefano Mottola and Uri Carsenty the following month. Lightcurve analysis gave a rotation period of 38.866 hours with a brightness amplitude of at least 0.34 magnitude (). While not being a slow rotator, Phoinix has a longer-than average period, especially for a large Jupiter trojan. Its relatively high brightness amplitude is indicative of a non-spherical, elongated shape. Diameter and albedo According to the surveys carried out by the Japanese Akari satellite, the NEOWISE mission of NASA's Wide-field Infrared Survey Explorer, and the Infrared Astronomical Satellite IRAS, Phoinix measures between 62.79 and 69.54 kilometers in diameter and its surface has an albedo between 0.049 and 0.059. The Collaborative Asteroid Lightcurve Link derives an albedo of 0.0540 and a diameter of 62.73 kilometers based on an absolute magnitude of 9.8. Naming This minor planet was named by the discoverer from Greek mythology after Phoenix (Phoinix), a wise Greek leader in the Trojan War, who raised Achilles and convinced him to join the campaign. The official naming citation was published by the Minor Planet Center on 28 April 1991 (). References External links Asteroid Lightcurve Database (LCDB), query form (info ) Dictionary of Minor Planet Names, Google books Discovery Circumstances: Numbered Minor Planets (1)-(5000) – Minor Planet Center 004543 Discoveries by Carolyn S. Shoemaker Minor planets named from Greek mythology Named minor planets 19890202
32861535
https://en.wikipedia.org/wiki/Dept.%20of%20Computer%20Science%2C%20University%20of%20Delhi
Dept. of Computer Science, University of Delhi
The Department of Computer Science, University of Delhi is a department in the University of Delhi under the Faculty of Mathematical Science, set up in 1981. Courses The department started the three years Master of Computer Applications (MCA) program in 1982, which was among the first such programs in India. The department started M.Sc. Computer Science course in 2004. Besides these the department has research interests in Computer Science and offers a Doctor of Philosophy (Ph.D.) program. The university conducts a postgraduate Diploma in Computer Applications (PGDCA) program through its constituent colleges. Emphasis is laid not only on the theoretical concepts but also on practical experience and industry interaction. Few classroom projects MCA Apart from classroom teaching, students take up case studies, presentations and small projects. Following are some projects/assignments taken up by the students: Implementation of Unix Shell Implementation of Chat Server. Simulation of machine language code and implementation of assembler. Simulation of the basic file system on Linux. Simulation of Sliding Window Protocols Go-Back N Protocol Selective Repeat Protocol. Simulation of a two-pass assembler. Projects designed, documented and coded using SDLC Share tracker system. Computerized health care system. Websites on tourism, online FIR, online book store, online examination, social networking, online shipping management system, digital library system. Research and implementation of cryptographic algorithms Design and implementation of new approach for searching in encrypted data using Bloom Filter. Analysis and implementation of security algorithms in Cloud Computing. Malware and keylogger design. Software and hardware implementation of Smart Home System. Misuse, detection and prevention of Advance Spamming techniques. Design and security analysis of chaotic encryption. Analysis of risks, techniques, and corporate usage of Web 2.0 technologies. Implementation of homomorphic encryption algorithms. Regional language encryption and translation. Implementation of elliptic curve cryptography. Design and implementation of self synchronizing stream ciphers. M.Sc. Computer Science As part of the curriculum students give presentations, group projects and programming assignments. The following are some of the projects/assignments taken up by the students: Implementation of robot task assignment given resources using MATLAB. Jade programming for agent communication. Implementation of DES encryption and decryption algorithm. Application of genetic algorithm in 8-queens problem. Implementation of K-means, FP-Tree, BIRCH and DBSCAN algorithm using C++. Generating all strong association rules from a set of given frequent item sets of transactions. Implementation of DBMS. Data preprocessing and KDD (Knowledge Discovery and Data mining) using WEKA and C4.5. Implementation of clustering techniques on output of fuzzy C-means algorithm as initial input using MATLAB. Simulation of Lexical Analyzer and Parser using C. Infrastructure The students of the department are affiliated to two libraries. The Departmental Library: is a reference library with over four thousand titles, in the field of Computer Science and IT and in related areas such as Electronics and Mathematics. The Central Science Library: is one of the largest science libraries in India. It was established in 1981, and has 2,20,000 volumes of books and periodicals. The website of CSL provides electronic subscription for 27,000 e-journals including IEEE, ACM, Springer journals and proceedings. Internet Connection All the labs, offices and faculty rooms of the Department are connected to the internet through the university intranet. Internet connectivity is provided using 4 switches through the university intranet. 24 port switch is used in LAN, providing internet to all systems in the laboratory, classrooms, seminar room and committee room. Delhi University Computer Centre Notable alumni Kiran Sethi - VP, Deutsche Bank, USA Pradeep Mathur - VP, Capgemini, UK Gulshan Kumar - Director, Alcatel-Lucent, India Ranjan Dhar - Director, Silicon Graphics, India Manish Madan - VP, Perot Systems, TSI, India Sachin Wadhwa - Head Operations, Mastech InfoTrellis Inc, USA Kumaran Sasikanthan -Country Head, AllSight Software, India References External links Official website Admissions Information Delhi University University departments in India
41838122
https://en.wikipedia.org/wiki/Michael%20A.%20Harrison
Michael A. Harrison
Michael A. Harrison is a computer scientist, in particular a pioneer in the area of formal languages. Biography Michael A. Harrison (born in Philadelphia, Pennsylvania, U.S.) studied electrical engineering and computing for BS and MS at the Case Institute of Technology, and then received a PhD from the University of Michigan in Communication Sciences. He was assistant professor from 1963 to 1966 at the University of Michigan, and then joined the faculty of the E.E. Dept at the University of California at Berkeley, where he was an associate professor from 1966 to 1971, and a full professor from 1971 to 1994. In the 1960s, he worked with Sheila Greibach, Gene Rose, Ed Spanier, and Joe Ullian in a research group formed and led by Seymour Ginsburg, dedicated to formal language theory and the foundations of Computer Science. The work that came out of this group distinguished Computer Science theory from other fields. It also brought the field of formal language theory to bear on programming language research. In 1975, he developed the HRU security model (named after its authors Harrison, Ruzzo, Ullman), an operating system level computer security model dealing with the integrity of access rights in the system. With his Ph.D. student Pehong Chen at Berkeley, he founded the "Gain Technology" company (acquired by Sybase in 1992). Currently, he is professor emeritus and also professor in the graduate school at Berkeley. Personal life Harrison is married to Susan L. Graham, the Pehong Chen Distinguished Professor Emerita in the Computer Science Division of the Department of Electrical Engineering and Computer Sciences at the University of California, Berkeley. References External links Publication list at acm.org Publication list at dblp.de Scientists from Philadelphia Theoretical computer scientists University of Michigan alumni UC Berkeley College of Engineering faculty Living people Year of birth missing (living people)
49860814
https://en.wikipedia.org/wiki/ACM-W
ACM-W
The Association for Computing Machinery's Council on Women in Computing (ACM-W) supports, celebrates, and advocates internationally for the full engagement of women in all aspects of the computing field, providing a wide range of programs and services to ACM members and working in the larger community to advance the contributions of technical women. ACM-W is an active organization with over 36,000 members. Celebrations of Women in Computing ACM-W sponsors annual celebrations focused on women in computing. ACM-W provides $3,000 for each celebration, and also raises and disburses corporate sponsorship. Each celebration organizing committee is responsible for additional fundraising within their conference area. ACM-W supports, celebrates, and advocates internationally for the full engagement of women in all aspects of the computing field, providing a wide range of programs and services to Association for Computing Machinery members and working in the larger community to advance the contributions of technical women. ACM-W Celebrations are regional conferences with global participants from industry, academia, and government. Celebration participation is growing and these events represent some of the largest gatherings of women in technology. The original Grace Hopper Celebration of Women in Computing was recognized by the US White House on their page "The Untold History of Women in Science and Technology" in the entry for United States Navy Rear Admiral Grace Hopper. In addition to this noteworthy beginning, the conferences have attracted the participation of technology notables including Anita Hill, Chan Zuckerberg Initiative co-founder Priscilla Chan, and Justine Cassell of Carnegie Mellon University, one of the top universities in Computer Science. The list describes the expansion of celebrations globally to include the largest gathering of women in computing in India. Canada Europe India Philippines Puerto Rico United Kingdom United States Chapters ACM-W has nearly 200 active chapters at schools around the world. The student chapters serve to increase recruitment and retention of women in computing fields at the university level and offer student activities and projects that aim to improve the working and learning environments for women in computing. Awards Starting in 2006, ACM-W has offered an annual Athena Lecturer Award to honor outstanding women researchers who have made fundamental contributions to computer science. ACM-W also offers an ACM-W Networking Award for active student chapters. Scholarships ACM-W provides support for women undergraduate and graduate students in Computer Science and related programs to attend research conferences. The ACM-W scholarships are offered for both intra-continental conference travel, and intercontinental conference travel. Scholarship applications are evaluated in six groups each year, to distribute awards across a range of conferences, including many annual ACM special interest group conferences such as SIGACCESS, SIGACT, SIGAI, SIGARCH, SIGCOMM, SIGCHI, SIGCSE, SIGDA, SIGECOM, SIGEVO, SIGGRAPH, SIGHPC, SIGIR, SIGITE, SIGMM, SIGMOBILE, SIGOPS, SIGPLAN, and SIGSOFT. Sponsors Sponsors of ACM-W services such as scholarships and regional celebrations include: Google Microsoft Research Oracle Academy Two Sigma Newsletter ACM-W publishes a monthly newsletter that highlights people, opportunities, accomplishments, and current issues associated with women in computing. The newsletter was started in 2008. Officers ACM-W officers include: Jody Tims, Chair Reyyan Ayfer, Vice Chair Melanie Wu, Treasurer Amelia Cole, Treasurer Arati Dixit, Standing Committee Chair Bushra Anjum, Standing Committee Chair Bettina Bair, Communications Committee Chair Sarah McRoberts, Communications Committee Chair Valerie Barr, Past Chair ACM-W regions and chairs are: Ruth G. Lennon, Europe Chair Heena Timani, India Chair Monica McGill, North America Jacqueline Tate, Asia Pacific Hong Gao, China ACM-W Standing Committees and Special Projects include: Viviana Bono, ACM-W Scholarships Pamela Wisniewski, ACM / ACM-W Awards Rising Star Rachelle Hippler, Professional Chapters Priya Chawla, Next Gen ACM-W Communications Committee members: Jennifer Goodall, ACM-W Connections Newsletter Editor See also Association for Computing Machinery CRA-W: Committee on the Status of Women in Computing Research List of organizations for women in science National Center for Women & Information Technology Women in computing References External links Celebrations of Women in Computing Women in Computing Oral History Collection Anita Borg Institute Association for Computing Machinery Information technology organizations Organizations for women in science and technology Women in computing Computer science education
6415506
https://en.wikipedia.org/wiki/Anil%20Neerukonda%20Institute%20of%20Technology%20and%20Sciences
Anil Neerukonda Institute of Technology and Sciences
Anil Neerukonda Institute of Technology and Sciences (ANITS), was established in the Academic Year 2001–02 with the approval of the All India Council for Technology Education (AICTE), New Delhi and the Government of Andhra Pradesh and is affiliated to Andhra University (AU), Visakhapatnam. All the eligible courses are Accredited by NBA in the year 2013. It is given Permanent Affiliation in the year 2010 by Andhra University. ANITS received autonomous status in the year 2015. The Institute is accredited by NAAC with A grade and valid up to 9 Dec 2019 with A grade. The institute CGPA is 3.01. Programmes Offered BTech Chemical Engineering Civil Engineering Computer Science & Engineering Computer Science & Engineering  (Data Science) Computer Science & Engineering (Artificial Intelligence  & Machine Learning) Electrical & Electronics Engineering Electronics & Communication Engineering Information Technology MTech M.Tech. (Control Systems) M.Tech.(Computer Science & Technology) M.Tech. (Communication Systems) M.Tech. (Machine Design) M.Tech. (Soil Mechanics) M.Tech. (Biotechnology) Universities and colleges in Visakhapatnam Colleges affiliated to Andhra University Educational institutions established in 2001 2001 establishments in Andhra Pradesh
308007
https://en.wikipedia.org/wiki/ETA%20Systems
ETA Systems
ETA Systems was a supercomputer company spun off from Control Data Corporation (CDC) in the early 1980s in order to regain a footing in the supercomputer business. They successfully delivered the ETA-10, but lost money continually while doing so. CDC management eventually gave up and folded the company. Historical development Seymour Cray left CDC in the early 1970s when they refused to continue funding of his CDC 8600 project. Instead they continued with the CDC STAR-100 while Cray went off to build the Cray-1. Cray's machine was much faster than the STAR, and soon CDC found itself pushed out of the supercomputing market. William Norris was convinced the only way to regain a foothold would be to spin off a division that would be free from management prodding. In order to regain some of the small-team flexibility that seemed essential to progress in the field, ETA was created in 1983 with the mandate to build a 10 GFLOPS machine by 1986. In April 1989 CDC decided to shut down the ETA operation and keep a bare-bones continuation effort alive at CDC. At shutdown, 7 liquid-cooled and 27 air-cooled machines had been sold. At this point ETA had the best price/performance ratio of any supercomputer on the market, and its initial software problems appeared to be finally sorted out. Nevertheless, shortly thereafter CDC exited the supercomputer market entirely, giving away remaining ETA machines free to high schools through the SuperQuest computer science competition. Products ETA had only one product, the ETA-10. It was a derivative of the CDC Cyber 205 supercomputer, and deliberately kept compatibility with it. Like the Cyber 205, the ETA-10 did not use vector registers as in the Cray machines, but instead used pipelined memory operations to a high-bandwidth main memory. The basic layout was a shared-memory multiprocessor with up to 8 CPUs, each capable of 4 double-precision or 8 single-precision operations per clock cycle, and up to 18 I/O processors. The main reason for the ETA-10's speed was the use of liquid nitrogen (LN2) cooling in some models to cool the CPUs. Even though it was based on then-current CMOS technologies, the low temperature allowed the CPUs to operate with a ~7 ns cycle time, so a fully loaded ETA-10 was capable of about 9.1 GFLOPS. The design goal had been 10 GFLOPS, so the design was technically a failure. Two LN2-cooled models were designated ETA-10E and ETA-10G. Two slower, lower-cost air-cooled versions, the ETA-10Q and ETA-10P (code named "Piper") were also marketed. The planned successor to the ETA-10 was the 30 GFLOPS ETA-30. Software Software for the ETA-10 line was initially regarded as a disaster. When CDC and ETA first designed the ETA architecture, they made the conscious decision not to merely port the CDC VSOS operating system from the existing CDC Cyber 205. It was felt by both the vendor, and the existing customer base (who wrongly believed that their vendor knew best), that a new OS needed to be written to extract the best performance from the hardware. When the first ETA-10 E initially shipped in 1986 there was no operating system for the machines. Programs had to be loaded one at a time from an attached Apollo Computer workstation, run, and then the supercomputer rebooted to run the next program. At the time Unix was making major inroads into the supercomputing fields, but ETA decided to write their own EOS operating system, which wasn't ready when the first machines were delivered in late 1986 and early 1987. An operating system based on UNIX System V became available in 1988, as a result of porting work done under contract by the Canadian firm HCR Corporation, at which point it looked like the machine might finally succeed. Many sites that had refused to pay for their machines due to the low quality of EOS found ETA's UNIX completely usable and were willing to accept delivery. ETA's demise was not based solely on operating system choice or existence. The Fortran compiler (ftn200) had not changed significantly from the CDC205. This compiler retained vendor-specific programming performance features (known as the Q8* subroutine calls) in an era when supercomputer users were realizing the necessity of source code portability between architectures. Additionally, the compiler optimizations were not keeping up with existing technology as shown by the Japanese supercomputer vendors such as NEC, or at Cray Research and the newer minisupercomputer makers. In general, computer hardware manufacturers prior and up to that period tended to be weak on software. Libraries and available commercial and non-commercial (soon to be called open-source) applications help an installed base of machines. CDC was relatively weak in this area. Naming the Firm Neil Lincoln, chief architect, asserts that ETA is not an acronym, and otherwise means nothing, not even the well known acronym of "estimated time of arrival." In point of fact, he says that he would not have named the firm that so as to dissociate it from, ETA, the Basque separatist group, or the Eta, the Japanese social minority (especially as Japan was considered a market). According to one of the CPU designers at ETA, Neil told a story that his sons actually came up with the name: apparently a linotype machine has characters arranged in the order of frequency used in the English language and the first three letters were used. Another theory asserts the name was chosen based on the frequency of English alphabet letters due in part to the then current popularity of Douglas Hofstadter's 1980 book Gödel, Escher, Bach which used ETAOIN, etc., to capitalize on popularity and current hip-ness. A third, completely plausible theory, is that ETA was a successor name to the much earlier Engineering Research Associates ERA: Engineering Technology Associates. Norris and others have denied this. See also Thomas Jefferson High School for Science and Technology won one of the donated ETA-10 supercomputers EOS was the initial operating system developed by ETA Systems Unix was the later, more popular operating system References Further reading Thorndyke, Lloyd (1994), "The Demise of ETA Systems", in Karyn R. Ames; Alan Brenner (eds.), Frontiers of Supercomputing II: A National Reassessment, pp. 489–496. A personal account of ETA Systems' history by its CEO. External links Computer history TNO location The Hague/Waalsdorp: the ETA10-P The ETA Saga American companies established in 1983 American companies disestablished in 1989 Computer companies established in 1983 Computer companies disestablished in 1989 Control Data Corporation Defunct computer companies of the United States History of electronic engineering
3467400
https://en.wikipedia.org/wiki/Star%20Trader
Star Trader
Star Trader is a 1974 video game and an early example of the space trading genre. The game involves a single player moving from star to star on a map of the galaxy, buying and selling quantities of six types of merchandise in order to make money. The game was developed by Dave Kaufman for computers in 1973, and its BASIC source code was printed in the January 1974 issue of the People's Computer Company Newsletter. It was reprinted in the 1977 book What to Do After You Hit Return. The game was the inspiration for the multiplayer Trade Wars series, beginning in 1984, and is thought to be the series the antecedent to much of the space trading genre. Gameplay Star Trader is a single-player space trading game, in which the player trades resources between star systems in order to make money. The game presents a star map of the galaxy in which the player moves about and makes money from trading and establishing trading routes. The player travels from node to node on the star map buying and selling six types of merchandise: uranium, metals, gems, software, heavy equipment, and medicine. The game's interface is text-only. Development Star Trader was written by Dave Kaufman in the BASIC programming language. The source code to the game was published in the People's Computer Company Newsletter in volume 2, issue 3 in January 1974. The concept for the game is seemingly based on Isaac Asimov's Foundation series of novels. Legacy In 1977, the game's code was reprinted in What to Do After You Hit Return. Star Trader was used as the inspiration for the first game of the Trade Wars series of multiplayer space trading games in 1984, making it the ancestor of many subsequent space trader games, including Eve Online, the Wing Commander Privateer series, and Elite series. Several unrelated but similar space trading games have been released under the name Star Trader. One such game was released by Bug Byte Software in 1984 for the ZX Spectrum and Commodore 64, and was one of the games included with the Softaid compilation. Another was written by S. J. Singer in 1984 using Altair Basic, and modified by John Zaitseff for Microsoft Basic under the CP/M-80 operating system in 1988. Completely rewritten versions for CP/M-80, CP/M-86, MS-DOS, Microsoft Windows 3.1 and Linux/Unix followed, with the latest release for Linux and Unix occurring in November 2019. All versions by John Zaitseff have been released or relicensed under the GNU General Public License v3. References External links The People's Computer Company Alumni Pages People's Computer Company Alumni and History site. Standard HP BASIC Listing #1 Star Trader Game Setup Module BASIC programming language listing. Standard HP BASIC Listing #2 Star Trader Game Main Module BASIC programming language listing. Star Trader Tribute Page Site describing the history of Star Trader. History Of Trade Wars Variants Timeline of the history of Trade Wars variants. Star Traders Versions of Star Traders for Linux, as well as for CP/M-80, CP/M-86, MS-DOS, Windows 3.1 and Commodore 64. 1974 video games Space trading and combat simulators Public-domain software with source code Video games developed in the United States
19358156
https://en.wikipedia.org/wiki/Pharmacosmos
Pharmacosmos
Pharmacosmos is a pharmaceutical company specialized in treatment of iron deficiency anemia. Company profile Pharmacosmos was founded in Denmark in 1965 by Henry Marinus Christensen, M.D., Ph.D. Operations Pharmacosmos' corporate headquarters, manufacturing facilities and research laboratories are located in Holbaek, Denmark (Scandinavia). The company has its own marketing, sales and services operations in Denmark and US, UK, Ireland, Germany, Sweden, Norway and China. In other geographies, the company partners with other pharmaceutical companies responsible for local marketing, sales and distribution of its pharmaceuticals brands. The company does not disclose its revenues. Manufacturing facilities Pharmacosmos' manufacturing facility is approved according to current Good Manufacturing Practice (GMP) by the European Medicines Agency (EMA) and by the U.S. Food and Drug Administration (FDA). Research and development The company has an ongoing clinical development program aimed at improving treatment options for patients suffering from iron deficiency anemia. In 2010, Pharmacosmos' clinical program included three international clinical studies focused on treatment of iron deficiency anemic patients within nephrology, gastroenterology and hematology/oncology. Products Monofer Monofer is a treatment for iron deficiency anemia. The generic name of Monofer is iron isomaltoside 1000, developed by Pharmacosmos. Monofer was approved in December 2009 in 22 European countries for treatment of iron deficiency anemia in patients with varying underlying conditions. Monofer is approved with a wide dose range and no test dose requirement, allowing iron repletion in a single IV dose if a Total Dose Infusion is given in certain clinical situations. This has to be given in a hospital due to increased risks of reaction. If the patient requires a dose over 20 mg/kg of idealised body weight a second dose has to be given after waiting a week. Cosmofer Cosmofer is a treatment for iron deficiency anemia. The generic name of Cosmofer is low molecular weight iron dextran. Cosmofer was approved in 2001 and is available in more than 45 countries worldwide. It has a lower incidence of dextran-induced analphylaxis than older dextran-based products, although still requires a test dose. Uniferon Uniferon is for the treatment and prevention of iron deficiency anemia in animals. The generic name for Uniferon is iron dextran. Its approved by health authorities in Europe, Asia and in the U.S. by Food and Drug Administration (FDA). Dextrans Pharmacosmos is one of the world leaders within production and fractionation of carbohydrates in particular Dextran. Pharmacosmos has a unique water-based technology and production platform for polymer fractioning and derivative synthesis. See also List of Danish companies List of pharmaceutical companies References Pharmaceutical companies of Denmark Veterinary medicine companies
1403586
https://en.wikipedia.org/wiki/List%20of%20University%20of%20California%2C%20Los%20Angeles%20people
List of University of California, Los Angeles people
This is a list of notable present and former faculty, staff, and students of the University of California, Los Angeles (UCLA). Notable alumni Nobel laureates Ralph Bunche – recipient of the 1950 Nobel Peace Prize Richard F. Heck – recipient of the 2010 Nobel Prize in Chemistry Robert Bruce Merrifield – recipient of the 1984 Nobel Prize in Chemistry Elinor Ostrom – recipient of the 2009 Nobel Memorial Prize in Economic Sciences Ardem Patapoutian - recipient of the 2021 Nobel Prize in Medicine Randy Schekman – recipient of the 2013 Nobel Prize in Medicine Glenn T. Seaborg – recipient of the 1951 Nobel Prize in Chemistry William F. Sharpe – recipient of the 1990 Nobel Memorial Prize in Economic Sciences Andrea Ghez - recipient of the 2020 Nobel Prize in Physics Academia, science and technology Arts and literature Amy Adler – artist Luis Aguilar-Monsalve – writer and educator Sara Kathryn Arledge – artist Catherine Asaro – Nebula Award-winning science-fiction novelist Glenna Avila – artist James Robert Baker – novelist Gary Baseman – artist Edith Baumann – abstract artist Rosa Beltrán – writer, lecturer and academic Guy Bennett – writer, translator and educator Susan Berman – author and screenwriter Stan Bitters – sculptor Justina Blakeney – designer and author Slater Bradley – artist Kenneth Wayne Bushnell – artist and educator Jan Butterfield – art writer and educator Vija Celmins – artist Judy Chicago – artist and educator Vicky A. Clark – curator Coleman Collins – artist Jennifer Dalton – artist Agnes de Mille – dancer and choreographer Jacques Ehrmann – literary theorist Dan Eldon – photojournalist Warren Farrell – educator, gender equality activist and author of The Myth of Male Power Alyce Frank – artist Martin Friedman – former director of Walker Art Center Charles Garabedian – artist Kelly Grovier – poet and literary critic Kim Gruenenfelder – author Sam Harris – writer Florence Parry Heide – author of children's literature Juan Felipe Herrera – professor, 21st U.S. poet laureate Gilah Yelin Hirsch – artist Diane Johnson – novelist Jane Jin Kaisen – artist Craig Kauffman – artist Jonathan Kellerman – Edgar Award-winning novelist and psychologist Annie Lapin – painter Mitchell Landsberg – journalist Gaylord Larsen – mystery writer Russell Leong – author and philosopher of Asian-American studies Linda Levi – artist Dave McNary – entertainment journalist Edward Meshekoff – artist Meleko Mokgosi – artist Ed Moses – artist Alexandra Nechita – painter John D. Nesbitt – writer and educator Tameka Norris – artist Flo Perkins – glass artist Raymond Pettibon – visual artist, known for creating the cover art for punk-rock band Black Flag's albums Jenelle Porter – art curator and author Jason Rhoades – artist Kay Ryan – poet and educator; U.S. Poet Laureate (2008–2010); MacArthur Fellow (2011) Betye Saar – artist Ben Sakoguchi – artist Shizu Saldamando – artist Sarah Seager – artist Cindy Shih – artist Jan Stussy (1921–1990) – artist Wu Tsang – artist Billie Tsien – architect, Barack Obama Presidential Center Harry Turtledove – Hugo Award and Nebula Award-winning science-fiction novelist Barbara Brooks Wallace – award-winning children's author, including two Edgar Awards and a William Allen White Children's Book Award * Emma Walton Hamilton (UCLA Lab School) – actress, author of children's book Idelle Weber – artist Antoine Wilson – novelist Jan Wurm – artist Richard Wyatt Jr. – artist Film, television and theater Athletics Hall of Famers in major team sports National Baseball Hall of Fame Jackie Robinson (inducted in 1962) Pro Football Hall of Fame Bob Waterfield (inducted in 1965) Tom Fears (inducted in 1970) Jimmy Johnson (inducted in 1994) Troy Aikman (inducted in 2006) Jonathan Ogden (inducted in 2013) Kenny Easley (inducted in 2017) Naismith Memorial Basketball Hall of Fame Baseball Basketball American football / Gridiron Golf John Merrick – professional golfer Corey Pavin – professional golfer, 1995 U.S. Open champion, 2010 Ryder Cup captain Tom Pernice, Jr. – professional golfer Monte Scheinblum – 1992 U.S. National and World Long Drive Champion Duffy Waldorf – professional golfer Gymnastics Association football / soccer Tennis Arthur Ashe (1943–1993) – tennis player (ranked as high as # 1) and social activist; Wimbledon champion and member of the International Tennis Hall of Fame Jimmy Connors (born 1952) – tennis player (ranked as high as # 1); two-time Wimbledon champion and member of the International Tennis Hall of Fame Herbert Flam (1928–1980) – tennis player (ranked as high as # 4) Zack Fleishman (born 1980) – professional tennis player Allen Fox (born 1939) – tennis player (ranked as high as # 4) and coach Mike Franks (born 1936) – professional tennis player Justin Gimelstob (born 1977) – professional tennis player Anita Kanter (born 1933) – tennis player ranked in world top 10 Steve Krulevitz (born 1951) – professional tennis player Kimberly Po – professional tennis player Brian Teacher – professional tennis player (ranked as high as # 7); Australian Open champion; and coach Eliot Teltscher (born 1959) – professional tennis player (ranked as high as # 6) Track and field Evelyn Ashford – Olympic track and field athlete and multiple gold medalist Ato Boldon – Olympic track and field athlete 1997 200 meter World Champion and four-time Olympic Sprint medalist Gail Devers – track and field runner; multiple Olympic gold medalist Danny Everett – Olympic bronze medalist in track and field Millard Hampton – track and field athlete, gold (4X100 relay) and silver (200 meters) medalist in the 1976 Montreal Olympics Dawn Harper – 2008 Olympics 100m Hurdles gold medalist Joanna Hayes – Olympic gold medalist track and field 100 m hurdles record holder Monique Henderson – track and field runner; Olympic gold medalist in 4 × 400 m relay Rafer Johnson – several-time world-record holder in the decathlon, and gold medalist at the 1960 Summer Olympics Florence Griffith Joyner – Olympic gold medalist and world record holder in 100 meter race Jackie Joyner-Kersee – track and field athlete, multiple Olympic gold medalist and world record holder in the heptathlon Meb Keflezighi – Olympic silver medalist, NCAA championships and New York City Marathon winner Steve Lewis – Olympic gold medalist in track and field Andre Phillips – track and field athlete, 400 meter hurdle gold medalist in the 1988 Seoul Olympics Mike Powell – former track and field athlete, current coach and holder of the long jump world record * Yang Chuan-kwang (also known as C.K. Yang) – former world-record holder in the decathlon, silver medalist in the decathlon in the 1960 Summer Olympics; first man to score over 9,000 points (using the tables at the time) Kevin Young – Olympic gold medalist in track and field, current world record holder 400 meters hurdles Volleyball Karch Kiraly – volleyball player and coach; only person to win Olympic gold medals in both indoor and beach volleyball Holly McPeak – beach volleyball player and Olympic bronze medalist Elaine Youngs – beach volleyball player and Olympic bronze medalist Water polo James Ferguson – 1972 Olympic bronze medalist, USA Water Polo Hall of Fame Natalie Golda – water polo player; Olympian Adam Krikorian – water polo player and coach; won 14 national titles Jillian Kraus (born 1986) - water polo player Monte Nitzkowski – Olympic water polo coach and swimmer Josh Samuels (born 1991) - Olympic water polo player Jovan Vavic - former head coach of the USC men's and women's water polo teams Other Glenn Cowan (1952–2004) - table tennis player Lisa Fernandez – Olympic softball gold medalist Brian Goodell – swimmer; nine NCAA individual championships, two-time Olympic champion, and former world record-holder Tommy Kendall – race car driver and television analyst Erwin Klein (d. 1992) - table tennis player Dan Kutler – Olympic swimmer Michelle Kwan – world champion figure skater; record nine-time U.S. National Champion Ken Pavia – former sports agent, founder of MMAagents Sports Agency, and the former CEO of India's first MMA Promotion Super Fight League Dot Richardson – softball player, Olympic gold medalist Mark Schultz (attended), 3x NCAA Champion, Olympic and world champion wrestler Doug Shaffer – platform diving, U.S. National champion, NCAA Diver of the Year, Head coach at UCLA, Minnesota and LSU Tim Thackrey – US National Team and Pan Am Games gold medalist Business and law Nancy J. Adler – Professor of Organizational Behavior and Samuel Bronfman Chair in Management at McGill University Eugene Anderson – attorney Shawn Holley – member of O. J. Simpson murder case defense team Fred D. Anderson – former CFO of Apple Computer John Edward Anderson – president of Topa Equities, Ltd.; namesake of UCLA Anderson School of Management Tom Anderson – founder of MySpace Nancy Austin – management consultant and author of The Assertive Woman Stephen F. Bollenbach – CEO of Hilton Hotels Corporation Saul Brandman (1925–2008) – garment manufacturer Bernard Briskin – co-founder and Chairman of Gelson's Markets Janice Rogers Brown – judge for the D.C. Circuit Court of Appeals Vincent Bugliosi – attorney and writer Michael Burry – hedge fund manager Frieda Rapoport Caplan – Entrepreneur in Specialty Produce Cormac J. Carney – United States federal judge Phil Carter – attorney, writer, and U.S. Army adviser in Iraq Morgan Chu – attorney, intellectual property expert Marcia Clark – attorney, lead prosecutor in O. J. Simpson murder case Johnnie Cochran – attorney Jeff Cohen – entertainment lawyer best known for work as child actor in The Goonies (1985) Lynn Compton – former judge for the California Court of Appeals and served as a commissioned officer with E Company, 2nd Battalion, 506th Parachute Infantry Regiment, in the 101st Airborne Division of the United States Army Roger Curtis – founder of Associated Electrics Keith Fink – attorney Laurence D. Fink – CEO and Chairman of BlackRock Dolly Gee – U.S. District Court judge Bill Gross – co-founder of PIMCO; philatelist Vinita Gupta – first Indian-origin woman to take her company public Horace Hahn – assisted Justice Robert H. Jackson as an interrogator in the prosecution of Nazi war criminals at the Nuremberg Trials Sam Hamadeh – co-founder, Vault.com John W. Henry – money manager and principal owner of the Boston Red Sox and Liverpool F.C. Nita Ing – chairman of Continental Engineering Corporation and Taiwan High Speed Rail William R. Johnson – chief executive officer, H.J. Heinz Company Alex Kozinski – judge for the Ninth Circuit Court of Appeals Ryan Lee – hedge fund manager and radio commentator Hardy McLain (born 1952) – hedge fund manager; managing partner of CVC Capital Partners Billy G. Mills – Los Angeles City Council member, 1963–74, Superior Court judge thereafter Marvin Mitchelson – attorney Irwin Molasky – real estate entrepreneur and early developer of Las Vegas Ezri Namvar – former founder and chairman of Namco Capital Group Dorothy W. Nelson – senior judge for the Ninth Circuit Court of Appeals Michael Newdow – plaintiff in Supreme Court case that challenged the constitutionality of the Pledge of Allegiance Robert C. O'Brien, United States National Security Advisor Michael Ovitz – Hollywood power broker and former president of the Walt Disney Company Robert O. Peterson – founder of the Jack in the Box restaurant chain Daniel Petrocelli – attorney Harry Pregerson – judge for the Ninth Circuit Court of Appeals Donald Prell – venture capitalist, author and futurist Subramaniam Ramadorai – chief executive officer and managing director, Tata Consultancy Services Jennifer Rodgers, former United States Attorney for the Southern District of New York and CNN legal analyst Nobutada Saji – chief executive officer, Suntory Robert Shapiro – attorney, part of defense team in O. J. Simpson murder case Sanford C. Sigoloff – businessman and philanthropist Stacey Snider – president of DreamWorks Ronald Sugar – chief executive officer, Northrop Grumman Jay Sures – co-president of United Talent Agency Edward Tabash – Constitutional attorney specializing in church and state issues; Board of Directors for the Center for Inquiry Robert Mitsuhiro Takasugi – federal judge A. Wallace Tashima – judge for the Ninth Circuit Court of Appeals Steven F. Udvar-Hazy – founder, chairman and CEO of ILFC Kim McLane Wardlaw – judge for the Ninth Circuit Court of Appeals Surangel Whipps Jr. - President of the Republic of Palau (2021 - Present) Don Yee – NFL sports agent James Yenbamroong – space entrepreneur and founder of Mu Space Ken Ziffren (J.D. 1965) – entertainment attorney, L.A.'s film czar (2014–present) Music Jenni Alpert – singer-songwriter Amir Bizzle – singer-songwriter Sara Bareilles – Grammy Award-winning singer-songwriter and pianist Jan Berry – singer-songwriter; member of the rock-and-roll duo Jan & Dean Jeff Blue – music producer for various labels; vice president of Warner California's artists-and-repertoire division Alison Brown – Grammy Award-winning banjo player John Cage – composer; student of Schoenberg Don Davis – film-score composer, including the science-fiction action film franchise The Matrix trilogy (1999, 2003 and 2003) Brad Delson – guitarist; lead guitarist and founding member of the Grammy Award-winning rock band Linkin Park Ryan Dusick – drummer, member of the Grammy Award-winning pop-rock band Maroon 5 Doriot Anthony Dwyer – principal flautist, Boston Symphony Orchestra Blake McIver Ewing − singer, pianist, actor John Fahey – experimental guitarist Jill Gibson – singer-songwriter, photographer, painter and sculptor Greg Ginn – guitarist and singer-songwriter; guitarist of the punk-rock band Black Flag Kim Gordon – musician; member of the alternative-rock band Sonic Youth Greg Graffin – singer-songwriter; lead singer of the punk-rock band Bad Religion Conan Gray – singer-songwriter;youtuber Joshua Guerrero, operatic tenor Este Haim, — member of Grammy nominated sister band HAIM. Jake Heggie – opera composer, Dead Man Walking James Horner – Academy Award, Golden Globe Award, and Grammy Award-winning film-score composer Anthony Kiedis – singer-songwriter; lead vocalist of the alternative-rock band Red Hot Chili Peppers Jim Lindberg – singer-songwriter; lead singer of the punk-rock band Pennywise Jon MacLennan – session musician and author of Melodic Expressions: The Art of the Line (2012) Mickey Madden – bass guitarist of the Grammy Award-winning pop-rock band Maroon 5 Ron Mael – musician and songwriter; co-founder (with brother Russell Mael) and keyboardist of the pop-rock band Sparks Russell Mael – singer-songwriter; co-founder (with brother Ron Mael) and lead vocalist of the pop-rock band Sparks Ray Manzarek – co-founder and keyboardist of the rock band The Doors Maile Misajon – singer-songwriter; former member of the pop girl group Eden's Crush Jim Morrison – poet and singer-songwriter; co-founder and lead vocalist of rock band The Doors Randy Newman – composer, pianist and singer-songwriter; Academy Award, Emmy Award and Grammy Award-winning film-score composer (dropped out one semester short of a B.A. in music) NS Yoon-G, stage name of Christine Kim, aka Kim Yoonji – South Korean singer John Ondrasik – singer-songwriter; performs pop rock under the stage name Five for Fighting Mo Ostin – music executive, Chairman Emeritus of Warner Bros. Records Kira Roessler – musician and film and television dialogue editor; bass guitarist of the punk-rock band Black Flag; film and television work includes Primetime Emmy Award-winning dialogue editor for her work on the biographical television miniseries John Adams (2008) episode "Don't Tread On Me" Laura Roppé – singer-songwriter and writer; cancer survivor who wrote memoir Rocking the Pink: Finding Myself on the Other Side of Cancer * Seo Jung-kwon – Korean-American rapper under the stage name Tiger JK, and leader of hip hop group Drunken Tiger Andy Sturmer – singer-songwriter and drummer of Jellyfish, producer for Puffy AmiYumi, composer of theme songs for Ben 10 and Teen Titans Paul Tanner – member of the Glenn Miller Orchestra, inventor of the Electro-Theremin instrument David Tao – singer Brian Tyler – BAFTA-nominated film score composer, conductor and film producer; his compositions include scores for Iron Man 3, Thor: The Dark World, and Frank Herbert's Children of Dune Kamasi Washington – jazz saxophonist, composer, producer and bandleader John Williams – Academy Award, Emmy Award and Grammy Award-winning composer; conductor and pianist; compositions include scores for the space opera Star Wars film series La Monte Young – composer, leading figure in musical minimalism Politics Shahid Khaqan Abbasi – Prime Minister of Pakistan Farid Abboud – ambassador of Lebanon to United States Senu Abdul Rahman – former Member of the Malaysian Parliament, Malaysia's first Minister of Information Glenn M. Anderson – United States Representative from California (1969–1993) Patrick Argüello – Nicaraguan-American revolutionary Howard Berman – member of the U.S. House of Representatives Tom Bradley – Mayor of Los Angeles (1973–1993) Yvonne Braithwaite Burke – Los Angeles County Supervisor John Campbell – member of the U.S. House of Representatives Benjamin Cayetano – Governor of Hawaii (1994–2002) Judy Chu – first Chinese-American woman ever elected to the U.S. Congress James C. Corman – Los Angeles City Council member; member of the U.S. House of Representatives J. Curtis Counts – director, Federal Mediation and Conciliation Service Edmund D. Edelman – Los Angeles City Council member (1965–1974); Los Angeles County Board of Supervisors member (1975–1994) John Ehrlichman – assistant and counsel to the Richard M. Nixon presidential administration Elizabeth Emken – 2012 Republican U.S. Senate Candidate Robert C. Farrell – journalist; Los Angeles City Council member (1974–1991) Dean Florez – member of the California State Senate, student body president at UCLA Kirsten Gillibrand – U.S. Senator from New York H.R. Haldeman – Chief of Staff for the Richard M. Nixon presidential administration; a key figure in the Watergate scandal James Day Hodgson – former United States Secretary of Labor and Ambassador to Japan Andrei Iancu - Director of the United States Patent and Trademark Office Paul Koretz – former member of the California State Assembly; Los Angeles City Council member Sheila Kuehl – former member of the California State Senate, California State Assembly, and current Los Angeles County Board of Supervisors Jerry Lewis – member of the U.S. House of Representatives, Chairman of the U.S. House Committee on Appropriations Calum MacDonald – former Member of Parliament in the United Kingdom Roberto Madrazo – candidate for president of Mexico in the 2006 presidential elections Jim Matheson – member of the U.S. House of Representatives for Utah Tom McClintock – member of the U.S. House of Representatives David McReynolds – activist and socialist political candidate Lloyd Monserratt – California political and community leader Bill Morrow – member of the California State Senate Gordon L. Park – former member of the Wyoming House of Representatives Dennis Ross – U.S. diplomat to the Middle East Edward R. Roybal – member of the U.S. House of Representatives Brad Sherman – member of the U.S. House of Representatives Helen Singleton – civil rights activist and Freedom Rider William French Smith – former United States Attorney General Todd Spitzer – member of the California State Assembly William R. Steiger – director, U.S. Department of Health and Human Services's Office of Global Health Affairs in the George W. Bush administration Ted Stevens – former senator of Alaska and alumnus to Delta Kappa Epsilon Rick Tuttle – Freedom Rider and Los Angeles City Controller Peggy Stevenson – Los Angeles City Council member (1975–1985) Antonio Villaraigosa – Mayor of Los Angeles; former Speaker of the California Assembly Joel Wachs – Los Angeles City Council member (1970–2001); president of the Andy Warhol Foundation for the Visual Arts in New York City Mimi Walters – member of the U.S. House of Representatives Diane Watson – member of the U.S. House of Representatives Henry Waxman – member of the U.S. House of Representatives Shirley Weber – Secretary of State of California Harold Willens – Jewish American businessman, political donor, nuclear freeze activist Helena Wong – member of the Legislative Council of Hong Kong, lecturer at Hong Kong Polytechnic University Zev Yaroslavsky (born 1948) – Los Angeles City Council member; Los Angeles County Board of Supervisors member Ehsan Zaffar – author; faculty; senior advisor on civil right,– U.S. Dept. of Homeland Security Miscellaneous Rodney Alcala – convicted rapist and serial killer active from 1968 to 1979, aka the "Dating Game Killer" for his successful appearance on The Dating Game Gustavo Arellano – OC Weekly writer and author of the "¡Ask a Mexican!" column Jules Asner – model and television personality Tony Auth – Pulitzer Prize–winning political cartoonist Rudi Bakhtiar – national news anchor Tony Blankley – commentator on The McLaughlin Group Joseph Blatchford – third Director of the United States Peace Corps Barbara Branden – author, Who is Ayn Rand? Nathaniel Branden – psychologist, author, Who is Ayn Rand?, Psychology of Self-Esteem and Judgment Day Judge Joe Brown – television judge Linda Burhansstipanov – Cherokee Nation of Oklahoma member, public health educator and researcher focused on Native American cancer care and support William George Carr – Executive Secretary of the National Education Association, 1952–1967 Carlos Castaneda – anthropologist and writer Paul Colichman – founder of Here! cable TV network Allen Cunningham – professional poker player Iva Toguri D'Aquino – World War II radio propagandist, "Tokyo Rose" Giada De Laurentiis – Food Network Chef (Every Day Italian) Clifford B. Drake - Marine Corps Major general Chris "Jesus" Ferguson – World Series of Poker main event winner and poker professional Harvey J. Fields – Reform rabbi. Alice Taylor Gafford – artist Jonathan Gold – Pulitzer Prize-winning food critic Kelly Goto – User experience design researcher and author of "Web Redsign, Workflow that Works." Josh E. Gross – publisher of Beverly Hills Weekly Todd Harris – Republican strategist on Hardball with Chris Matthews Frank B. James – U.S. Air Force general Arthur Janov – psychologist, inventor of primal therapy Stephen Francis Jones – American architect known for high-end restaurant designs Kang Dong-suk – yachtsman, first Korean solo circumnavigator Jill Kinmont – educator, quadriplegic, alpine ski racer in the 1950s Ida B. Kinney – civil rights activist Ezra Klein – blogger; journalist, Vox.com Ralph Lazo – civil rights activist, only known non-spouse and non-Japanese American who voluntarily relocated to a World War II Japanese American internment camp Flora Lewis – journalist with The New York Times Carol Lin – national news anchor Laura Ling – journalist with Current TV, notable for her detainment in North Korea Bridget Marquardt – co-star of The Girls Next Door Nana Meriwether – Miss Maryland USA 2012, Miss USA 2012 Emily Mitchell – co-president of Anderson Wine Club Donn Moomaw – Presbyterian minister, member of the College Football Hall of Fame K. Patrick Okura – Japanese American psychologist and civil rights activist Zoltan Pali – architect Steve Parode – U. S. Navy Rear Admiral William R. Peers – U.S. Army Lt. General best known for leading the army's investigation of the My Lai incident Kelly Perdew – winner of The Apprentice Brian R. Price – author, editor, publisher, martial-arts instructor of the Italian school of swordsmanship, reconstructive armorer, and Dissertation Fellow in History at the University of North Texas Lila Rose – activist and president of Live Action Steve Sailer – paleoconservative blogger and journalist (VDARE, Taki's Magazine) James M. Seely – U.S. Navy Rear Admiral and acting Assistant Secretary of the Navy (Financial Management and Comptroller) from December 18, 1988 to January 1990 Ben Shapiro - conservative commentator for The Daily Wire Marcus Stern – Pulitzer Prize-winning journalist Alan S. Thompson – retired U.S. Navy Vice Admiral Daniel Thompson – inventor of the automatic bagel maker and the folding ping pong table Princess Ubol Ratana of Thailand Francis B. Wai – Medal of Honor recipient J. Warner Wallace – homicide detective and Christian apologist Stephen Worth – director of the ASIFA-Hollywood Animation Archive Project Prince Chatri Chalerm Yukol of Thailand Kelly, Hereditary Princess of Saxe-Coburg and Gotha Notable faculty Nobel laureates Paul D. Boyer – professor of chemistry; recipient of the 1997 Nobel Prize in Chemistry Donald Cram (1919–2001) – professor of chemistry; recipient of the 1987 Nobel Prize in Chemistry Louis J. Ignarro – professor of molecular and medical pharmacology; recipient of the 1998 Nobel Prize in Medicine Willard Libby (1908–1980) – professor of chemistry; recipient of the 1960 Nobel Prize in Chemistry Bertrand Russell (1872–1970) – mathematician and philosopher; recipient of the 1950 Nobel Prize in Literature Julian Schwinger (1918–1994) – professor of physics; recipient of the 1965 Nobel Prize in Physics Lloyd Shapley – professor of economics; recipient of the 2012 Nobel Prize in Economics Business Mark A.R. Kleiman – professor of public policy, noted expert on crime and drug policy William Ouchi – professor of management and best-selling author Law Medicine Michael S. Gottlieb – first physician to diagnose AIDS David Ho – AIDS researcher Howard Judd – menopause expert and medical researcher Martha Kirkpatrick (1925–2015) – clinical professor of psychiatry Kimberly J. Lee – reconstructive surgeon Linda Liau – W. Eugene Stern Chair of the Department of Neurosurgery Courtney Lyder – expert in gerontology; first black dean of the UCLA School of Nursing No-Hee Park, DMD, PhD – Dean, UCLA School of Dentistry and notable researcher of oral (head and neck) cancer and aging research Patrick Soon-Shiong – executive director, UCLA Wireless Health Institute Sarah Meeker Jensen – FAIA, Founder of Jensen Partners Healthcare Planning Politics Michael Dukakis – professor of policy studies, former governor of Massachusetts and 1988 presidential candidate Al Gore – visiting professor, 45th Vice President of the United States Larry Pressler – teacher and visiting fellow, former Senator from South Dakota Science and technology George O. Abell (1927–1983) – professor of astronomy Asad Ali Abidi – professor of electrical engineering; pioneer of CMOS RF circuits; member of the National Academy of Engineering M. C. Frank Chang – professor of electrical engineering; member of the National Academy of Engineering Alonzo Church – known for the lambda calculus used in computing Steven Clarke – professor of chemistry and biochemistry; pioneer in protein repair in aging (L-isoaspartyl methyltransferase) Vijay K. Dhir – dean of the UCLA Henry Samueli School of Engineering and Applied Science; professor of mechanical and aerospace engineering François Diederich – professor of chemistry David Eisenberg – professor of chemistry and biochemistry, and of biological chemistry; Director of the UCLA-DOE Institute for Genomics and Proteomics Sergio Ferrara – professor of physics; co-discovered supergravity in 1976 Rajit Gadh – professor of mechanical and aerospace engineering; founder and director of UCLA Smart Grid Energy Research Center and Wireless Internet for Mobile Enterprise Consortium William Gelbart – professor of chemistry and biochemistry Andrea Ghez – professor of astronomy; expert in the galactic center and adaptive optics; Crafoord Prize recipient Sheila Greibach – professor of computer science, known for the Greibach normal form A. M. Harun-ar-Rashid, physicist; member, Nobel Committee for Physics Kendall Houk – professor of chemistry Tatsuo Itoh – professor of electrical engineering; member of the National Academy of Engineering Michael E. Jung – professor of chemistry Richard Kaner – professor of chemistry Alan Kay – professor of computer science; Turing Award laureate Margaret Kivelson – professor of space physics; expert in planetary magnetospheres; member of the National Academy of Sciences Leonard Kleinrock – professor of computer science; Internet pioneer; recipient of the 2007 National Medal of Science William Scott Klug – associate professor of mechanical and aerospace engineering; killed in the 2016 UCLA shooting Raphael David Levine – professor of chemistry Tung Hua Lin – professor of civil and environmental engineering; designer of China's first twin-engine aircraft Seymour Lubetzky – professor of library and information science Donald A. Martin – professor of mathematics and philosophy Mildred Esther Mathias – professor of botany (1962 – 1974), eponym of the campus' Mildred E. Mathias Botanical Garden William V. Mayer – professor of zoology Carlo Montemagno – Associate Director, California Nanosystems Institute; Founding Department Chair, Department of Biomedical Engineering; Carol and Roy Doumani Professor of Biomedical Engineering (2001-2006); Father of Bionanotechnology Henry John Orchard – professor of electrical engineering; pioneer of the field of filter design Mangalore Anantha Pai – power engineer, Shanti Swarup Bhatnagar laureate Stott Parker - Professor of computer science Judea Pearl – professor of computer science; pioneer of Bayesian networks and the probabilistic approach to artificial intelligence; Turing Award laureate Roberto Peccei – professor of physics; former dean of the UCLA Division of Physical Sciences; Sakurai Prize recipient Theodore M. Porter – professor of history of science Abraham Robinson – professor of mathematics and philosophy Joseph Rudnick – professor of physics; former dean of the UCLA Division of Physical Sciences Amit Sahai - professor of computer science Lloyd Shapley – professor of mathematics; known for the Shapley value in game theory Elizabeth Stern – professor of epidemiology Ernst G. Straus – professor of mathematics Terence Tao – professor of mathematics; Fields Medalist in 2006; Crafoord Prize recipient Edward Wright – professor of astronomy; expert in cosmology and infrared astronomy; member of the National Academy of Sciences Jeffrey Zink – professor of chemistry and biochemistry Social science, arts and humanities Rogers Albritton – late professor of philosophy Perry Anderson - Marxist historian; professor emeritus of History and Sociology Joyce Appleby – U.S. historian; specialist in intellectual history and the legacy of liberalism Ann Bergren – professor of Greek Literature, winner in 1988 of the University's Distinguished Teaching Award. William Bodiford – professor of Japanese and Buddhist studies Tyler Burge, Professor, Fellow, American Academy of Arts and Sciences Kenny Burrell – professor of jazz studies; jazz guitarist and composer Rudolf Carnap – late professor of the philosophy of language Alonzo Church – pioneer in the philosophy of language and computer science James Smoot Coleman – Africanist; founded the UCLA African Studies Center Brian Copenhaver – emeritus historian of philosophy Denis Cosgrove – Alexander von Humboldt Professor of Geography Angela Davis – assistant professor of philosophy, fired in 1969 by the Board of Regents and California Governor Ronald Reagan for her membership in the Communist Party Jared Diamond – professor of geography and physiology, Pulitzer Prize-winning author of Guns, Germs and Steel: The Fate of Human Societies Keith Donnellan – late professor of philosophy Frederick Erickson – professor emeritus of educational anthropology Kit Fine – former professor of philosophy Philippa Foot – late professor of philosophy Steven Forness – Professor of Psychiatry and Biobehavioral Sciences Andrea Fraser – professor of interdisciplinary studio Saul Friedländer – European historian; specialist in Holocaust studies Lowell Gallagher – literary theorist James Gimzewski – physicist and nanotechnology pioneer Carlo Ginzburg – European historian; pioneer of microhistory Juan Gómez-Quiñones – U.S. historian; specialist in Chicano history Lev Hakak – Professor of Hebrew Language and Literature at UCLA N. Katherine Hayles – literary critic Barbara Herman – professor of philosophy Thomas Hines – architectural historian; professor emeritus James N. Hill (1934–1997) – processualist archaeologist Darnell Hunt (PhD UCLA) – Professor of Sociology and African American Studies, Dean of Social Sciences Neil Peter Jampolis – professor of theater Design; Tony Award Winning Designer; Director and Designer of Theater, Dance, and Opera Donald Kalish – late professor of philosophy Abraham Kaplan – late professor of philosophy David Kaplan – professor of the philosophy of language Edmond Keller – professor of political science; Africanist Harold Kelley (1921–2003) – professor of psychology; social psychologist Mark Kleiman – professor of public policy, expert on crime and drug policy Peter Kollock (1959–2009) – associate professor of sociology, specialist in collaboration and online participation in virtual communities Peter Ladefoged – professor of linguistics, specialist in phonetics Deborah Nadoolman Landis — professor of costume design; Oscar-nominated costume designer of Coming to America; founding director of the David C. Copley Center for Costume Design David Kellogg Lewis – former assistant professor of philosophy Barbara Kruger – professor of new genres, recipient of the Leone D’Or award from the Venice Biennale Ole Ivar Lovaas – professor of psychology, specialist in applied behavior analysis therapy for autism Michael Mann – professor of sociology; author of The Sources of Social Power volumes I and II Mwesa Isaiah Mapoma – Zambian musicologist Julián Marías – philosopher, opponent of Francisco Franco, author of History of Philosophy Thom Mayne – professor of architecture, architect, co-founder of firm named Morphosis Susan McClary – musicologist; prominent in the new-musicology movement; MacArthur Fellow; works have been translated into over twelve languages; wrote Feminine Endings: Music, Gender, and Sexuality Vasa Mihich – professor of design and media arts; artist and sculptor Richard Montague – late professor of philosophy Charles Moore – professor of architecture, author and architect, Beverly Hills Civic Center Richard Thacker Morris (1917–1981) – chairman of the sociology department, author Donald Neuen – professor of choral studies; conductor; apprentice of Robert Shaw Calvin Normore – professor of philosophy Karen Orren – professor of political science; noted for her work in American political development Catherine Opie – professor of photography and recipient of the Guggenheim Fellowship Terence Parsons – professor of philosophy John Perry – former professor of philosophy Lari Pittman – distinguished professor of painting Hans Reichenbach – late professor of philosophy Amy Richlin – professor in Department of Classics Teofilo Ruiz – European historian; specialist in medieval history Bertrand Russell – former professor of philosophy; taught as a guest lecturer for one year David Schaberg – Dean of Humanities Arnold Schoenberg – professor of music; composer Seana Shiffrin – professor of philosophy and law known for her work in legal and moral philosophy Josef von Sternberg – taught film aesthetics Shelley Taylor – professor of psychology; social psychologist Dominic Thomas – chair of the department of French and Francophone Studies at UCLA Eugen Weber – historian; author of Peasants Into Frenchmen Luc E. Weber – Rector Emeritus of the University of Geneva Dixon Wecter – professor of English (1939 to 1945). Louis Jolyon West – professor of psychiatry; specialist in brainwashing Gerald Wilson – professor of ethnomusicology, jazz studies; jazz composer, arranger and musician Eugene Victor Wolfenstein – professor of political science; author of Psychoanalytic-Marxism: Groundwork Roy Bin Wong – professor of history; pioneer in modern Chinese economic history Medha Yodh – professor of classical Indian dance John Zaller – political scientist; author of The Nature and Origins of Mass Opinion Amy Zegart – professor of public policy and U.S. intelligence analyst; author of Spying Blind Athletics Athletic directors Fred Cozens – Director of Physical Education and Athletics (1919–1942), first basketball (1919–1921) and football (1919) head coach Dan Guerrero, B.A. 1974 – athletic director (2002– ) Wilbur Johns, 1925 – athletic director (1948–1963), men's basketball head coach (1939–1948), basketball player J. D. Morgan – athletic director (1963–1979), head tennis coach (1949–1966), tennis player (1938–1941) Basketball coaches Steve Alford – men's basketball head coach (2013–2018) Frank Arnold – men's basketball assistant coach (1971–1975) Gene Bartow – men's basketball head coach (1975–1977) Larry Brown – men's basketball head coach (1979–1981), member of the Basketball Hall of Fame Tasha Butts – women's basketball assistant coach Nikki Caldwell – women's basketball head coach (2008–2011) Cori Close – women's basketball head coach (2011–present), women's basketball head coach (1993–1995) Mick Cronin – men's basketball head coach (2019–present) Denny Crum, 1958 – men's basketball assistant coach (1963–1971), player (1956–1958), member of the Basketball Hall of Fame Gary Cunningham – men's basketball head coach (1977–1979), basketball player (1960–1962) Donny Daniels – men's basketball assistant coach (2003–2010) Larry Farmer – men's basketball head coach (1981–1984), basketball player (1970–1973) Mark Gottfried – men's basketball assistant coach (1987–1995) Jim Harrick – men's basketball head coach (1988–1996) Walt Hazzard – men's basketball head coach (1984–1988), basketball player (1961–1964), NBA and Olympic player Jack Hirsch – men's basketball assistant coach (1984–1988), player (1961–1964) Brad Holland, B.A. 1979 – men's basketball assistant coach (1988–1992), player (1975–1979) Michael Holton – men's basketball assistant coach (1996–2001), player (1979–1983) Ben Howland – men's basketball head coach (2003–2013); 2006 Pac-10 Coach of the Year Chad Kammerer – men's basketball assistant coach Kerry Keating – men's basketball assistant coach (2003–2007) Steve Lavin – men's basketball head coach (1996–2003) Gerald Madkins – men's basketball assistant coach, player Philip Mathews – men's basketball assistant coach (2010–2013) Billie Moore – women's basketball head coach (1977–1993), member of the Basketball Hall of Fame Kevin O'Connor – men's basketball assistant coach (1979–1984) Kathy Olivier – women's basketball head coach (1993–2008), women's basketball head coach (1986–1993) Lorenzo Romar – men's basketball assistant coach (1992–1996) Jim Saia – men's basketball assistant coach (1996–2003) Kenny Washington – first women's basketball head coach (1974), basketball player (1963–1966) Greg White – men's basketball assistant coach (1995–1996) Sidney Wicks – men's basketball assistant coach (1984–1988), player (1968–1971) John Wooden – men's basketball head coach (1948–1975), won 10 NCAA championships, member of the Basketball Hall of Fame as both a player and coach Caddy Works – men's basketball head coach (1921–1939) Ernie Zeigler – men's basketball assistant coach (2003–2006) Football coaches Sal Alosi – strength and conditioning coordinator Dino Babers – assistant head coach (2004–2007) William F. Barnes – head football coach (1958–1964) Eric Bieniemy – running backs coach (2003–2005) Gary Blackney – assistant coach (1978–1979) Sam Boghosian – assistant coach (1957–1964), player (1952–1954) James J. Cline – head football coach (1923–1924) George W. Dickerson – interim head football coach (1958) Terry Donahue – head football coach (1976–1995), football player (1965–1966), member of the College Football Hall of Fame Karl Dorrell – head football coach (2003–2007), football player (1983–1986) Edwin C. Horrell – head football coach (1939–1944), member of the College Football Hall of Fame Mike Johnson – interim head coach (2011) Ed Kezirian – interim head football coach (2002), football player Adrian Klemm – run game coordinator and offensive line coach Bert LaBrucherie, 1929 – head football coach (1945–1948), football player (1926–1928) Demetrice Martin – defensive backs coach Noel Mazzone – offensive coordinator Jim L. Mora – head football coach (2012– ) Rick Neuheisel, B.A. 1984 – head football coach (2008–2011), football player (1980–1983) Kennedy Polamalu – running backs coach Tommy Prothro – head football coach (1965–1970), member of the College Football Hall of Fame Pepper Rodgers – head football coach (1971–1973) Henry Russell Sanders – head football coach (1949–1957) Lou Spanos – defensive coordinator (2012–2013) William H. Spaulding – head football coach (1925–1938) Bob Toledo – head football coach (1996–2002) Harry Trotter – head football coach (1920–1922), track coach (1919–1946) Jeff Ulbrich – assistant head coach Dick Vermeil – head football coach (1974–1975) DeWayne Walker – interim head football coach (2007) Eric Yarber – wide receivers coach Miscellaneous coaches Elvin C. Drake – head sports trainer, 1942–1972; head track and field coach, 1947–1964, winning the NCAA championship in 1956; coached decathletes Rafer Johnson and Yang Chuan-kwang during the 1960 Summer Olympics, in which they won the gold and silver medals Adam Krikorian – men's and women's water polo coach, won nine NCAA championships; assistant coach, won one NCAA championship; UCLA water polo player, won 1995 NCAA championship Al Scates – men's volleyball coach, won 19 NCAA championships John Smith – track and field coach, inventor of the drive phase and world record holder at event See also List of people from Los Angeles References University of California, Los Angeles people Los Angeles people
20209624
https://en.wikipedia.org/wiki/List%20of%20features%20removed%20in%20Windows%207
List of features removed in Windows 7
Windows 7 contains many new features. However, similarly to the transition from Windows XP to Windows Vista, certain capabilities and programs that are present in Windows Vista are no longer present in Windows 7 or have changed. The following is a list of features that were present in Windows Vista and earlier versions but were removed in Windows 7. Windows shell Windows Explorer Windows 7 removes the "Customize" tab for a folder if its properties are opened via Libraries. If the same folder is opened in Explorer without accessing it via Libraries, the Customize tab is available in its Properties. The address bar no longer lists the Public folder introduced in Windows Vista in its hierarchy; the user must manually travel to Public. The Registry settings for modified colors for compressed files, EFS-encrypted files, item hot-tracking (single click hover) etc. are ignored by the ItemsView control in Windows 7 Explorer. Running Explorer shell as administrator or even opening an Explorer window temporarily as administrator is not possible in Windows 7 without modifying permissions in the registry of system values because of a DCOM restriction. Control Panel views are limited to Small icons and Large icons (which is actually Tiles view). List view, Details view, Extra large icons view, Large icons view and Medium icons view with text at the bottom are not available. The ability to show Control Panel items in groups is also removed. IDeskBand interface and Explorer Bars are no longer supported in Windows Explorer but still supported in Internet Explorer. Some Control Panel settings, such as Add Hardware, Bluetooth, Game Controllers, Pen and Touch, People Near Me, Tablet PC Settings, Offline Files, and Scanners and Cameras, are not listed, even under All Control Panel Items. The 32-bit Speech applet is no longer accessible through the Control Panel. While some of these settings are accessible only from Category view, or by searching, they can be made accessible via All Control Panel Items with a modification to the registry. The slide show button from Windows Explorer starts the Windows Photo Viewer, which—unlike Windows Live Photo Gallery—does not support viewing slide shows with themes and rich transitions. It is not possible to run the 32-bit version of Windows Explorer as a file manager or as the shell in 64-bit editions of Windows 7. 64-bit editions of Windows Vista and Windows XP allowed executing the 32-bit shell/Windows Explorer and making it the default for compatibility with shell extensions. As a result, all 32-bit shell extensions are incompatible with 64-bit versions of Windows 7. Because of the Software Explorer being removed from Windows Defender, there is no user notification any more if UAC blocks a startup program which needs administrator privileges. Users were notified about blocked startup programs set to run as administrator in Windows Vista. An undocumented incompatible Items view control replaces the List view control used in Windows Explorer and Start Menu breaking applications that customized the Explorer list view control. The Item view also does not support custom positioning, custom ordering, or hyperlinks, which the Windows Vista list view did support. Because the control is undocumented, it is not possible for developers to disable auto arrange or customize views. Because ItemsView is not considered part of the Win32 common controls reusable by developers, the sound for the Select system event no longer plays in Explorer in Windows 7. The new items view control also ignores certain desktop.ini parameters like IconArea_Image and IconArea_Text preventing users from setting a background picture for folders. In Windows Vista, this feature was not built-in, however the list view control supported this if shell extensions or folder customization utilities which set the correct desktop.ini parameters and attributes were installed. Each toolbar or desk band is forced to be on its own row in Windows Explorer. Windows Explorer forces the RBBS_BREAK style for every band. When items are grouped by any criteria, multiple selection of groups by clicking the group header while holding down is not possible as could be done in Windows Vista. The entire folder tree view in the navigation pane cannot be collapsed vertically as could be done in Windows Explorer on Windows Vista. Full row selection cannot be turned off for details view as was possible in Windows Vista. The Share overlay icon for shared items in Explorer has been removed. When navigating to a folder from a library, it is not possible to view the folder's contents in the view the user has set or customized. The folder uses the library's view. The Remember each folder's view settings option to set Explorer to persist or not persist folder views and icon sizes per folder has been removed. It is not possible to set Windows Explorer to remember individual folder window sizes and positions. Each folder window shares the same size, and their position cascades as more windows are opened. Dotted lines that connect folders and subfolders do not appear as the option to turn off simple folder view is missing. When clicking on a folder in the folders pane, it isn't automatically expanded to show its subfolders as the option to turn on simple folder view is missing. The size of any selected item(s) is not shown on the status bar but is shown in the details pane just above the status bar. Free disk space is not shown on the status bar and the status bar is not visible by default. The details pane only shows the size of 15 selected items by default and viewing the total size of more than 15 selected items requires clicking "Show more details" each time the selection changes. When no items are selected in a folder, neither the details pane nor the status bar show the total size of files in the folder. Icons on the command bar in Windows Explorer and common application dialogs for Open/Save. Disabling Auto Arrange and Align to Grid is not available in Windows Explorer windows. As a result, users can no longer manually arrange items via drag and drop within a folder; this feature is only available on the desktop. The bar with column headers for sorting, grouping, filtering and stacking that appears for any view in Windows Vista Explorer only appears for Details view. Even in details view, the column headers only allow sorting and filtering; they no longer allow grouping and stacking. As a result, grouping must be done from the context menu in all views and sorting must also be done from the context menu in all other views except Details view. Filtering by columns can only be done in Details view. Automatic horizontal scrolling in the left pane is removed as well as the horizontal scrollbar to manually scroll. The Arrange By option (called Stack By in Windows Vista) is only available for libraries and folders included in libraries. In addition, users can only Stack By predefined categories depending on the library type. For example, users can only Stack by Folder, Album, Artist, Song, Genre, and Rating for the Music library, whereas in Vista, users could Stack By any properties that exist for a file (common examples would be bit rate, year, composer, play count, etc. for audio files). Changing the sort criteria when multiple items are selected causes all items to be de-selected. Selection of multiple items is not retained across back and forward navigation. Taskbar Tooltips for program icons only show the name of the app. They no longer show the comment field. 16-bit applications can no longer display their icon on the taskbar when running. The taskbar shows the icon for NTVDM.exe, which is the 32-bit host process for running 16-bit applications. Quick Launch was removed in favor of 'pinned' applications. It is possible to revert to Quick Launch by manually adding it as a custom toolbar referencing %AppData%\Microsoft\Internet Explorer\Quick Launch (check citation for how-to) if users prefer the taskbar layout from earlier versions of Windows, although Windows 7 and later have issues with Quick Launch getting automatically disabled and all its settings getting reset after a logoff or restart. Clicking a grouped application's taskbar button or thumbnail when it is active (in the foreground) does not minimize it. Only ungrouped application buttons minimize upon being clicked again. Icons have been removed from the tooltips for the notification area system icons such as Volume, Network, Power and the taskbar date and time. Floating Deskbands (toolbars) are no longer available. The feature was previously deprecated in Windows Vista; all deskbands can only be pinned to the taskbar. It is no longer possible to place toolbars on the other (non-taskbar) edges of the desktop. The ability to disable grouping (placing next to each other) similar taskbar buttons (for example, Explorer windows) is removed; users can still disable combining multiple taskbar buttons of the same type into one button. The configurable registry option TaskbarGroupSize, which allowed fine-grained control over taskbar grouping under previous Windows versions is not supported. In Windows 7, users can only configure to either combine buttons, not combine or combine when the taskbar is full. The number of combined taskbar windows is not calculated and shown next to the combined buttons. Network activity animation on Network Connections icon in the notification area cannot be turned on. The Network icon only indicates the type of network (wired or wireless) and Internet connectivity. The ability to turn off taskbar window previews (thumbnails) while still using Aero, as was configurable in Windows Vista. It is possible to turn off taskbar thumbnails when two or more windows are combined by setting a registry value, though. The ability to disable the taskbar's "always on top" setting is removed. When grouping is disabled, the ability to manage multiple taskbar items using multiselect (Ctrl+click) to tile, cascade, minimize or close the selected group of windows has been removed. Only two power plans, one of them always being Balanced, and the other being "Power saver" by default, are available to choose from the taskbar. The "High Performance" power plan is hidden by default in the Control Panel. When the taskbar is vertical, there cannot be multiple columns of icons. Start menu Applications can no longer be excluded from appearing in the Start Menu's Most Frequently Used program list using the NoStartPage registry value as was possible in Windows XP and Windows Vista. Programs started from outside the Start Menu such as from the Desktop, the Run dialog or from folders no longer show up in the Start Menu's Frequently Used Programs list, unless launched at least once from the Start Menu. Classic Start menu is removed. This feature can be added by third-party programs, like Classic Shell. Certain features that were present only in the Classic Start menu are not available, such as expanding the menu columns instead of scrolling, expanding folders by hovering instead of clicking, opening folders by double clicking, launching multiple programs by holding down Shift key while clicking, and creating expandable shortcuts to folders by dragging them to the Start menu. Dynamically pinning default Internet browser and e-mail client software programs on Start menu is no longer possible. Programs can still be, however, manually pinned to the Start menu in Windows 7. Internet Explorer Favorites and History are no longer grouped under a separate header in the Windows 7 Start Menu as they were in the Windows Vista Start Menu. They are shown under the Files group. Windows Search features Advanced search builder UI (Search Pane). Due to the removal of this UI, it is not possible to change the location of where to search in the same Explorer window unless at least one search has already been performed, after which Explorer shows a button to search a "Custom" scope. For saved searches (stacked views), changing the scope of the search (location for example) is no longer possible because of this removal. The option to always search only file names has been removed from Folder Options, Search tab. The Windows Search protocol handler addin to index network paths is incompatible with Windows 7. The Windows Search service needs to be installed and running on the remote computers to search them. Pre-defined saved searches in the Searches folder ("Recent Documents", "Recent E-mail", "Recent Music", "Recent Pictures and Video", "Recently Changed" and "Shared by Me") have been removed. Other shell features The Snap To mouse pointer option to move the pointer automatically to the default button in a dialog is broken on many re-designed system and application dialog boxes and windows in Windows 7. The mouse pointer simply does not move or snap to the default button in several dialogs which are re-designed. For security reasons, AutoPlay will only be supported on optical media, such as CDs and DVDs. (Later Microsoft released KB971029 for earlier OSes to make AutoPlay Windows 7-like.) Detection of the discontinued HD DVD format in AutoPlay. Properties such as target command line with switches, start in path, icon and compatibility mode cannot be changed for game shortcuts in Games Explorer. The Windows command-line interpreter (cmd.exe) no longer displays the RAM usage (maximum: 190 MB at 9999×9999) when selecting large buffer grids. The “About” or “Information” windows in preinstalled Windows applications (such as Microsoft Paint or Windows Calculator) as well as the winver dialog no longer displays a given system's available RAM capacity, a detail that was previously included in the dialog of NT-based Windows versions since Windows NT 3.5. Personalization In the Personalization control panel, it is not possible to save changes to an existing theme. Changes made after choosing a theme must be saved again and the original theme deleted to prevent duplication. In previous versions of Windows, a theme file could be overwritten with the modified theme. The Aurora, Windows Energy and Windows Logo screensavers and the wallpapers that shipped with Windows Vista were removed. The 3D-Bronze, 3D-White, Conductor, Dinosaur, Hands 1, Hands 2, Variations and Windows Animated cursor schemes are removed and no longer available. ClearType cannot be turned off entirely in order for the user interface font, Segoe UI, to maintain optimal design for certain shell components and Windows Explorer. Parts of the user interface (such as the start menu and Explorer) still use ClearType regardless of setting. Icon size setting in the Advanced Appearance settings opened via Personalization dialog no longer affects icon sizes in Windows Explorer and on the Desktop. Windows Media Player features The ability to lock the player while in full-screen mode using a 4-digit PIN has been removed. The option to adjust the bit rate when burning data CDs has been removed. Windows Media Player's taskbar-integrated Mini-player has been removed. The thumbnail preview which replaces this lacks volume control and a progress bar. The ability to start the miniplayer only for certain files based on specific text in their file names was also removed. Enhancements are only accessible from Now Playing view in a floating window. They do not dock to the Now Playing window, and do not get restored when Windows Media Player is restarted. Even when manually restored, their position does not get saved. Left/Right balance control has been removed from the graphic equalizer and is no longer possible. Several player preferences are not saved and restored upon restarting the player. The playlist pane in Now Playing view is not shown automatically. Enhancements do not get restored when Windows Media Player is restarted. Even when manually restored, the previous position of the Now Playing window and enhancements is not retained. The context menu entry "Find In Library" which allowed locating the Now Playing song in the library was removed. Advanced Tag Editor, Color Chooser, Media Link for E-mail, and Party Mode features have been removed. The ability to add and show static lyrics and synchronized lyrics has been removed. Recently added Auto playlist is not included by default. Total time for CDs and playlists is rounded up to the nearest minute, i.e. seconds are no longer shown. Internet Explorer features In Internet Explorer 8, the previous session can no longer be automatically restored the next time. The user must remember to manually open the session the next time. Internet Explorer 8 on any Windows version does not support inline AutoComplete in its address bar. Rating and description of Internet Explorer favorites can no longer be edited from the Details pane in Windows 7 Explorer. Applications replaced by Windows Live counterparts Windows Photo Gallery, Windows Movie Maker, and Windows Mail have been removed, in favor of downloading respective Windows Live counterparts, after they were discontinued in favor of third-party e-mail clients, photo and video software. Windows Live Mail and Windows Live Movie Maker do not include all of the features of Windows Mail and Windows Movie Maker respectively. Windows Live Mail does not support HTML source editing, scripted stationery and the ability to disable the splash screen. The Windows Mail gadget for Windows SideShow does not ship with Windows 7. Windows Live Movie Maker does not include a proper timeline with audio edits and narration, title overlays, filter compatibility settings and other advanced options, cannot export to DV-AVI, and drops the transitions and effects from Windows Movie Maker. The web filtering and activity reporting functionality from the in-box parental controls feature. Windows Live Family Safety replaces the web filtering functionality but its user interface for filtering and activity reporting is web-based instead of the native UI used by Windows Vista's parental controls and requires logging into a Windows Live ID (now Microsoft account). After, that services was discontinued in favor of third-party parental control software. Windows 7 retained time, game and program restrictions. Windows Calendar has also been removed, in favor of Windows Live Mail's calendar. Windows Live Mail's integrated calendar supports events but not tasks, does not support subscribing, publishing, importing and exporting iCalendar files. Windows Live Mail's integrated calendar also lacks a native events reminder. Events are reminded through e-mail instead. Other Windows applications and features To preserve the digital signature of a Windows Installer (MSI) file during uninstall, embedded cabinets are no longer automatically stripped off by Windows Installer 5.0 to save disk space. As a result, MSIs cached in the %Windir%\Installer directory take more disk space than earlier Windows versions with no option to disable the caching. Transient Multimon Manager (TMM), a Windows Vista feature to automatically detect hot-Plug and Play displays as well as configure and save the multi-monitor setup is removed in favor of Win+P. The Windows 7 On-screen keyboard removes the ability to change or specify the font as a result of which keyboard layouts of fonts which use Dingbats and Unicode characters in place of alphabetic characters can no longer be viewed. The Windows 7 On-screen keyboard removes direct access to function keys. Function keys are now accessed using the 'Fn' key. Windows Remote Assistance does not support file transfer and clipboard sharing in Windows 7. The Offline Files cache can be relocated, however, an existing cache with previously cached files cannot be moved without resetting it as was possible in Windows Vista using Windows Easy Transfer or in Windows XP/2000 using Cachemov.exe The DFS Replication Service included in Windows Vista for peer-to-peer DFS Replication service groups, has also been removed. The 3D visuals from WinSAT benchmark have been removed. Some features have been removed from Windows Defender such as the Software Explorer, configuration of security agents and notifications of program activities or when running startup programs as administrator. Removable Storage Manager (RSM) has been removed, so applications that depend on it—such as NTBackup or the NTBackup Restore Tool—cannot back up to physical or virtual tape drives. Windows Meeting Space, and the InkBall game have been removed. Windows Ultimate Extras for the Ultimate edition which included Windows DreamScene, Microsoft Tinker, and the Hold' Em poker game among others, was removed. Microsoft had initially announced that they will re-offer Tinker and Hold 'Em in the near future. On December 15, only Tinker was made available for free as a Games for Windows – Live downloadable game which requires signing into Windows Live ID. Microsoft Agent 2.0 technology was removed, breaking compatibility with Microsoft Agent applications. A separate installation package has been provided by Microsoft "due to customer feedback." The sidebar for Desktop Gadgets was removed, although gadgets can align to any side of the screen like they did in Vista. The Contacts, Notes and Stocks Desktop Gadgets have been removed. The Notes gadget has been replaced by Sticky Notes, allowing for both text and ink-based notes. An upgrade from Windows Vista to Windows 7 retains the user's notes from the gadget and transfers them to the Sticky Notes program. The import, export, drag and drop and voice note features in earlier versions of Sticky Notes have been removed. The Windows Firewall Control Panel does not allow configuring port-based filtering although the Windows Firewall with Advanced Security MMC-based snap-in allows this. Filtering policies in Group Policy to show only a specific operating system or application is no longer available. Software Restriction Policies no longer support multiple levels of trust such as "basic user" (only block or allow are still supported); this functionality has been superseded by User Account Control and AppLocker. The common font selector dialog box does not show the type of font (OpenType, TrueType, Symbol font, etc.) Windows Registry Reflection for x64 editions replaced with merged 32/64-bit keys. File backup to a network share (non-system image backup) is only available in Windows 7 Professional and above whereas it was included in Windows Vista Home Premium. Windows Import Video, a feature in Windows Vista which allowed one to import live or recorded video from a digital video camera and save it to the hard disk, has been removed. The option in Windows Vista to send search queries (keywords) of searches performed in the Control Panel category view to Microsoft has been removed in Windows 7. See also Windows Server 2008 R2 References Windows 7 Windows 7 Microsoft lists Computing-related lists
34362463
https://en.wikipedia.org/wiki/Check%20MK
Check MK
Checkmk is software developed in Python and C++ for IT Infrastructure monitoring. It is used for the monitoring of servers, applications, networks, cloud infrastructures (public, private, hybrid), containers, storage, databases and environment sensors. Checkmk is available in three editions: an open source edition ("Checkmk Raw Edition – CRE"), a commercial enterprise edition ("Checkmk Enterprise Edition – CEE") and a commercial edition for managed services providers ("Checkmk Managed Services Edition – CME"). These Checkmk-Editions are available for a range of platforms, in particular for various versions of Debian, Ubuntu, SLES and Red Hat / CentOS, and also as a Docker Image. In addition, physical appliances of various sizes as well as a virtual appliance are offered to simplify the administration of the underlying operating system through a graphical user interface and to enable high-availability solutions. The agents used by Checkmk to collect data are available for 11 platforms, including Windows. History Checkmk originated in 2008 as an Agent-substituting shell script for Inetd, and was published in April 2009 under GPL. It was initially based on Nagios, and extended this with a number of new components. The open source edition (Checkmk Raw Edition) also continues to be based on the Nagios-core, and bundles this with additional open source components into a complete system. Over many years Checkmk's commercial editions have evolved into a self-contained monitoring system – one that has replaced all of the essential Nagios components with its own – including its very own monitoring core. The majority of the developments for the commercial editions, in particular all plug-ins, are also available into the Checkmk Raw Edition. While in the past Checkmk was designed for monitoring large and heterogeneous on-premise environments, from version 1.5+ (1.5p12) it also supports the monitoring of AWS, Azure, Docker and Kubernetes services. Checkmk is being developed by tribe29 GmbH in Munich Germany, which until 16.04.2019 operated under the name of Mathias Kettner GmbH. Together with the company name change, the product name "Check_MK" was also changed to "Checkmk". tribe29 GmbH follows an open core business model. The open source edition is available under different open source licenses – mostly GPLv2, while large parts of the commercial editions run under the proprietary "Checkmk Enterprise License". The Product Checkmk combines three types of IT monitoring: Status-based monitoring, which (via thresholds) records the "health" of a device or application. Metric-based monitoring that enables the recording and analysis of time series graphs using a HTML5-based graphing system. An integration with Grafana is available as well. Log-based and event-based monitoring, in which key events can be filtered out and actions can be triggered based on these events. In order to ensure a very broad monitoring, Checkmk currently has 1700+ plug-ins in each edition – all of which are licensed under GPLv2. These plug-ins are maintained as part of the product and are regularly supplemented with additional plug-ins or extensions. Connecting existing legacy Nagios plug-ins is possible as well. To simplify setup and operation, all components of Checkmk are delivered fully integrated. A rule-based 1:n configuration, as well as a high degree of automation significantly accelerate workflows. This includes: Auto-discovery of hosts (where applicable) Auto-discovery of services Automated configuration of plug-ins via preconfigured thresholds and rules Automated agent updates (a CEE feature) Automatic and dynamic configuration that enables the monitoring of volatile services with a lifespan of just a few seconds, such as in the Kubernetes environment (starting from CEE v1.6) Automated discovery of tags and labels from sources such as Kubernetes, AWS and Azure (starting from CEE v1.6) In addition, there are also playbooks for the use of configuration and deployment tools such as Ansible or Salt. Checkmk is often used in very large distributed environments where a high number of sites (e.g. 300 locations of Faurecia) and/or well over 100.000 devices (e.g. Edeka) are monitored. This is possible, among other things, because Checkmk's microcore consumes much less CPU resources than, for example Nagios, and therefore offers a significantly higher performance on the same hardware. Furthermore the non-persistent data is stored in-memory in RAM which significantly improves the access time. Components The Monitoring Core ("Checkmk Microcore - CMC") The commercial editions of Checkmk use their own monitoring core, written in C++. This has a much higher performance than the Nagios core. In addition, as of version 1.6 it allows the dynamic recording of objects with a short lifespan, such as Containers. This is possible because - in contrast to the Nagios core - the Checkmk Microcore does not require a reboot when making configuration changes. The open source version "Checkmk Raw Edition" currently still uses the Nagios core. Configuration & Check Engine Checkmk offers self-contained service discovery and settings generation. Checkmk uses its own method when carrying out the checks. During the test period each host is contacted only once. The test results are transmitted to the monitoring core as passive checks. This significantly improves the performance on the monitoring server, as well as on the hosts being monitored. Checkmk uses different methods to access the data in the target systems. These include agents installed on the target system, "special agents" running on the monitoring server and communicating with the API of the target system, the SNMP API for monitoring, for example, network devices and printers, and HTTP/TCP protocols to communicate with web and internet services. By default, Checkmk follows the "pull principle", i.e. the data is explicitly queried by the monitoring system to quickly identify when a system suddenly fails and does not respond to a "pull". As an alternative, however, a "push" can be configured with which the system transfers its data directly to Checkmk or to an intermediate host. Data Interface ("Livestatus") Livestatus is the main interface in Checkmk. It provides live access to all data from the monitored hosts and services. The data is fetched directly from the RAM, which avoids slow hard disk access and gives fast access to the information without overloading the system too much. Access is done via a simple protocol and it is possible from all programming languages without requiring a special library. Web-GUI ("Multisite") Multisite is Checkmk’s web GUI. In addition to having a quick page layout, it offers user-definable views and dashboards, distributed monitoring by integrating multiple monitoring instances via Livestatus, integration of NagVis, an integrated LDAP connection, access to status data via web services, and much more. Dashboards and views can be differentiated for various users or groups of users, for example vSphere-specific views for VMware admins. The web GUI is currently available both in German and English. Setup Checkmk is completely administrable via the browser via its Setup module. This includes managing users, roles, groups, time periods, and more. Permissions can be granted in a granular way using a role concept. Existing role-based access controls (LDAP, AD) can be used for this. Checkmk works rule-based, so that the configuration remains intuitive even in complex environments, and the necessary effort is low. Automatic discovery and configuration, as well as the automatic agent update further accelerate the configuration process. An HTTP API can also be used to integrate CMDBs for accelerated configuration. Alert System Several notification channels can be set up and configured with different rules for each user. For example, emails can be triggered at any time of the day, but notifications via SMS are sent only for important issues during on-call hours. The notifications can be set for all or for specific teams, e.g. notify only the storage admins about a failed hard drive. Duplicate notifications are grouped together so that no user is notified twice through a particular channel. Furthermore, users can configure their own notifications themselves. In distributed environments alerts can be managed centrally. For detected issues, actions can be triggered automatically (alarm control) via scripts. Checkmk includes integrations to email and SMS gateways as well as to communication and IT service-management solutions such as Slack , Jira , PagerDuty , OpsGenie , VictorOps and ServiceNow. Business Intelligence The BI module is integrated into the graphical user interface. It aggregates the overall status of business processes, their dependency on complex applications and IT infrastructure elements from many individual hosts and services in a rule-based manner. It can also be used to represent applications made up of microservices, which in turn consist of Kubernetes pods and deployments. In addition, worst-case scenarios can be simulated in real time and historical data can be analyzed to understand the causes of performance degradation. Event Console The Event Console integrates the processing of log messages and SNMP traps into the monitoring. It is configured via a flexible set of rules, and decides whether incoming messages are to be discarded or how they are to be classified. It can count, correlate, expect messages, rewrite messages, and more. Similar entries can be grouped into a single event (e.g. multiple failed logins) to keep track of events. It also has a built-in syslog daemon that receives messages directly on port 514, and an SNMP trap receiver that receives traps on port 162. Metrics Graphing The commercial Checkmk editions use their own metric and graphing system. This allows time series metrics to be analysed over long intervals using interactive HTML5 graphs. The maximum resolution is one second. Data can be imported from a variety of data sources and metrics formats (JSON, XML, SNMP etc.) and stored on the disk of a long-term data storage device. Alternatively, Graphite or InfluxDB can be connected via an export interface. From CEE version 1.5p16 there is also a plug-in available for integrating data directly from Checkmk into Grafana for visualization purposes. The Checkmk Raw Edition currently uses PNP4Nagios as its graphing system. Reporting Reporting enables the direct delivery of PDF reports, ad-hoc or automatically, at regular intervals. It includes the availability analysis in which the history of the states over any desired time period can be provided with a click. Availability calculations can exclude unmonitored times, adjust the resolution, or ignore short intervals. In addition to the availability calculations, reporting also includes SLA reporting in which complex SLAs can be monitored. The reporting is only available in the commercial versions of Checkmk. Hardware/Software Inventory The hardware/software inventory can be used, for example, to monitor hardware and software changes, to verify the presence of installed security updates, and to update static data with dynamic parameters (for example, updating the current disk usage statistics based on monitoring data). The Configuration Management Database (CMDB) i-doit has a deep integration that enables the exchange of CMDB data with monitoring data. See also Comparison of network monitoring systems References External links Computer monitoring with the Open Monitoring Distribution (Kelvin Vanderlip, 2012-03-01) Using the Open Monitoring Distribution(Nagios) to Monitor Complex Hardware/Software Systems (Joe VanAndel, 2012-03-29) Free network management software Free software programmed in Python Nagios System monitors Network management Systems management
12223381
https://en.wikipedia.org/wiki/VistA%20imaging
VistA imaging
VistA Imaging is an FDA-listed Image Management system used in the Department of Veterans Affairs healthcare facilities nationwide. It is one of the most widely used image management systems in routine healthcare use, and is used to manage many different varieties of images associated with a patient's medical record. The system was started as a research project by Ruth Dayhoff in 1986 and was formally launched in 1991. Hardware requirements The VistA Imaging System uses hardware components to provide short- and long-term storage. It takes advantage of network servers for storage. It uses a DICOM gateway system to communicate with commercial Picture Archiving and Communication Systems (PACS) and modalities such as CT, MR, and Computed Radiography (x-ray) devices for image capture. It utilizes a background processor for moving the images to the proper storage device and for managing storage space. Types of data managed The system not only manages radiologic images, but also is able to capture and manage EKGs, pathology images, gastroenterology (endoscopic) images, laparoscopic images, scanned paperwork, or essentially any type of health care image. Integration with Electronic Health Record systems VistA Imaging is currently integrated into the VistA EMR (electronic medical record) system used nationwide in Department of Veterans Affairs hospitals. This integration is able to provide increased efficiency of retrieval of images. It has also been used as a separate software package and can be used with EHRs other than VistA. VistA Imaging now connects to a nationwide backbone that allows clinicians to access the 350 million images stored in the VA system via Remote Image View software. The VA has developed interfaces for more than 250 medical devices in VistA Imaging, the images from which can be accessed through the desktop VistA Imaging Viewer. The Department of Defense will use the VistA Imaging Viewer to enhance its own EHR. Usage in a National Network of Healthcare Records As part of the US national mandate to co-ordinate care between Department of Defense and the VA, VistA Imaging is forming a cornerstone of the effort to exchange medical imagery between the two systems. “When soldiers come back from Iraq and Afghanistan and eventually enter the VA system, images will be able to move from DOD to VA seamlessly." Eventually, DOD and VA should be able to share all image file types from all sites. Additional enhancements to VistA Imaging include development of a central archive for all VA images (whether acquired through VistA or a commercial system) and new indexing and search capabilities. Availability The software for VistA Imaging has been made available through the Freedom of Information Act so that it is in the public domain. Due to its designation as a medical device, however, it can not be designated as free open source software and therefore can not be altered or implemented without FDA approval. Although it can be used in healthcare facilities that are outside the Department of Veterans Affairs, this is possible only if the proprietary modules that have been integrated into it are also licensed and implementation is registered with the FDA. This has effectively limited its use to government institutions who have licensed the proprietary modules. The source code can be downloaded from the OSEHRA VistA-M.git tree. Proprietary modules required VistA Imaging uses proprietary modules not in the public domain. This makes its public domain use limited. Information retrieval after a natural disaster The VistA Imaging system was robust enough to be restored after Hurricane Katrina damaged the data facility at the New Orleans VA. This type of backup proved superior to a paper record system. References External links United States Department of Veterans Affairs Health care software Medical imaging Nuclear medicine Endoscopy Medical equipment Public-domain software DICOM software
41048790
https://en.wikipedia.org/wiki/Federation%20University%20Australia
Federation University Australia
Federation University Australia (Fed Uni) is a public, dual-sector university based in Ballarat in Victoria, Australia. The university also has campuses in Ararat, Horsham, Stawell, Churchill, Berwick, and Brisbane, as well as online technical and further education (TAFE) courses and Horsham's Higher Education Nursing program. Federation University is the fourth oldest tertiary education institution in Australia, having begun under predecessor institutions in 1870, during the Victorian gold rush. With the merger between the University of Ballarat and Monash University's Gippsland campus in 2013, the university changed its name to Federation University from 2014. History 1870–2013 Tertiary education at Ballarat began in 1870, making it Australia's fourth oldest tertiary institution. 2014–present On 6 September 2013, the Victorian Parliament passed legislation to establish Federation University Australia, The name change officially began in 2014. The then Vice-Chancellor justified the name change as an attempt to broaden the reach of the university nationally and internationally, and in fairness to the campuses outside Ballarat. The name was also the preferred name within the university, compared to 'State University of Victoria', 'Eureka University', 'Robert Menzies University' or 'Vida Goldstein University'. Buildings and architecture The former Ballarat Gaol, a maximum security prison that operated from 1862 until 1965, was located on the site of the university's School of Mines (SMB) campus, at the southern end of Lydiard Street. The area is known for being a well preserved Victorian era street. While the prison was mostly demolished in the 1960s, the old prison walls, gate and guard towers, as well as the residences of the governor and warden, still exist. The Collaborative Research Centre in Australian History (CRCAH) is located in the governor's residence of the Old Ballarat Gaol. The CRCAH director's office was once the bedroom of Bella Guerin, who in 1883 became the first woman to graduate from an Australian university. The campus also includes the old School of Mines buildings. Academia Undergraduate studies Students can undertake undergraduate degrees across a wide range of study areas, which are: Humanities and social sciences Engineering Business Science and mathematics Education and early childhood Nursing, midwifery and paramedicine Psychology Performing arts Visual arts Information technology Occupational health and safety Sport, outdoor and physical education TAFE Research Researchers – academics and post-graduate students – undertake work within various centres, as well as within the disciplines. The research priority areas of the university are information forensics and security, transformative and preventative health, dynamic landscapes, history and heritage, and improving policy and practice in VET. The research centres are: Australian Retirement Research Institute (ARRI) Centre for Biopsychosocial and eHealth Research and Innovation (CBeRI) Centre for eResearch and Digital Innovation (CeRDI) Centre for Gippsland Studies (CGS) Centre for Informatics and Applied Optimisation (CIAO) Centre for Multimedia Computing, Communications, and Artificial Intelligence Research (MCCAIR) Collaborative Research Centre in Australian History (CRCAH) Geotechnical and Hydrogeological Engineering Research Group (GHERG) Researching Adult and Vocational Education (RAVE) Water Research Network Visiting Friends and Relatives Research There are also research facilities at Ballarat Technology Park, the Gippsland Enterprise Centre and Nanya Station in rural NSW. Rankings In 2017, FedUni was ranked within the top 20% of Australian universities in humanities for teaching quality, learner engagement, learning resources, skills development, and student support. Nationally, the university is ranked 31 in Australia (out of 40) and 1526 in the world. Across the university, for postgraduate research the university is rated above the national average by former students with 78.8% satisfaction. For undergraduate studies, the university is rated above national average by former students with 83.3% satisfaction. Historically, the former University of Ballarat achieved a maximum five-star rating for teaching quality in the Good Universities Guide consecutively from 2010 to 2014. Federation University has been ranked four-stars for graduate placement by the Good Universities Guide, but is not ranked in international university guides. This placed the university in the top tier of Australian regional universities. Student demographics and engagement In 2017, 80% of undergraduate students study full-time and on campus, which is unique for a regional university, and 35% of students are international students. Facilities Campuses In addition to the following campuses, the university also had joint-degree programmes with international colleges, including PLK Vicwood KT Chong Sixth Form College in Hong Kong. Ballarat Camp Street Campus – located in central Ballarat, this campus houses the Arts Academy. The campus consists of the Old General Post Office Building, the Old Courthouse, and several newer buildings which were completed in 2002. SMB Campus – located in central Ballarat and incorporates the original School of Mines Ballarat and the Old Ballarat Gaol. The campus offers training from Certificate level through to Advanced Diploma and Degree level study. Online learning: Federation University offers many courses through online study. This consists of online reading materials, online virtual classes, online quizzes and weekly course activities. Western Victorian campuses Horsham Campus – offers TAFE courses and a Higher Education course in Nursing. Mount Helen campus Located in Mount Helen, 10 km south of Ballarat. The university's largest campus, it has three residences, Peter Lalor South Hall, Peter Lalor North Hall, and Bella Guerin Hall. Its programs include the School of Education and Arts; School of Health Sciences; School of Science, Information Technology and Engineering; and the Business School. Gippsland campus The Gippsland Campus is located in the township of Churchill in the foothills of the Strzelecki Ranges. The campus is home to over 2,500 students and approximately 400 staff The campus was formerly Monash University, Gippsland campus, but became part of Federation University Australia on 1 January 2014. Berwick campus The Berwick Campus is located south east of the Melbourne city centre. It was transitioning itself from the Monash University, Berwick campus in 2017 and completed its transition in early 2018. The exact location of the building is 100 Clyde Road Berwick VIC 3806. The university has four buildings naming 901, 902, 903, and 930 as well as additional buildings for on-campus living. Nursing has the highest enrolment rate for Berwick campus, also focusing on education courses, IT and Psychology. Brisbane campus The Brisbane campus, situated in the Brisbane city area, offers a wide range of undergraduate and postgraduate programs in Information Technology, Business and Allied Health. Technology park The university has a technology park with the mission to facilitate the development of technology-based companies or companies that benefit from the technological resources of the university. The following organisations operate in the park. Ambulance Victoria Conservation Volunteers Country Fire Authority Emergency Services Telecommunications Authority (ESTA) Global Innovation Centre Greenhill Enterprise Centre IBM South East Asia IBM Regional Software Solutions Centre ID Research State Revenue Office More than 1350 people are employed by tenants at the technology park and approximately half of those holding Federation University Australia qualifications. Recently IBM decided to expand its workforce with the construction of a new $10 million building on the park. Research institutes and centres Institute for Regional and Rural Research and Innovation Geotechnical and Hydrogeological Engineering Research Group Collaborative Research Centre in Australian History Notable alumni Aunty Donna, absurdist sketch comedy troupe William Baragwanath, Geologist Phillip Bellingham, winter olympian Steve Bracks, former Premier of Victoria Sandy Blythe, wheelchair basketball player Dr Cyril P. Callister, inventor of Vegemite Jacqueline Dark, opera singer David Davies (artist) Keith Hamilton, former Minister for Agriculture and former Minister for Aboriginal Affairs William Roy Hodgson, human rights diplomat Will Longstaff, artist Brad McEwan, Ten Network sports reporter Kiran Mazumdar-Shaw, businesswoman and associate of Biocon Steve Moneghetti, Olympic marathon runner Richard W Richards, physicist and Antarctic explorer Libby Tanner, actress Marcus Wills, artist Dai Xulong, President of Economic Diplomacy Center of China Ministry of Commerce CAFIEC Martin Andanar, press secretary of the Philippines under Duterte administration See also List of universities in Australia Notes References External links Federation University Australia – official website Australian vocational education and training providers Universities in Victoria (Australia) Ballarat Educational institutions established in 1994 1994 establishments in Australia
36827
https://en.wikipedia.org/wiki/Mutual%20exclusion
Mutual exclusion
In computer science, mutual exclusion is a property of concurrency control, which is instituted for the purpose of preventing race conditions. It is the requirement that one thread of execution never enters a critical section while a concurrent thread of execution is already accessing critical section, which refers to an interval of time during which a thread of execution accesses a shared resource, such as [Shared data objects, shared resources, shared memory]. The shared resource is a data object, which two or more concurrent threads are trying to modify (where two concurrent read operations are permitted but, no two concurrent write operations or one read and one write are permitted, since it leads to data inconsistency). Mutual exclusion algorithm ensures that if a process is already performing write operation on a data object [critical section] no other process/thread is allowed to access/modify the same object until the first process has finished writing upon the data object [critical section] and released the object for other processes to read and write upon. The requirement of mutual exclusion was first identified and solved by Edsger W. Dijkstra in his seminal 1965 paper "Solution of a problem in concurrent programming control", which is credited as the first topic in the study of concurrent algorithms. A simple example of why mutual exclusion is important in practice can be visualized using a singly linked list of four items, where the second and third are to be removed. The removal of a node that sits between 2 other nodes is performed by changing the next pointer of the previous node to point to the next node (in other words, if node is being removed, then the next pointer of node is changed to point to node , thereby removing from the linked list any reference to node ). When such a linked list is being shared between multiple threads of execution, two threads of execution may attempt to remove two different nodes simultaneously, one thread of execution changing the next pointer of node to point to node , while another thread of execution changes the next pointer of node to point to node . Although both removal operations complete successfully, the desired state of the linked list is not achieved: node remains in the list, because the next pointer of node points to node . This problem (called a race condition) can be avoided by using the requirement of mutual exclusion to ensure that simultaneous updates to the same part of the list cannot occur. The term mutual exclusion is also used in reference to the simultaneous writing of a memory address by one thread while the aforementioned memory address is being manipulated or read by one or more other threads. Problem description The problem which mutual exclusion addresses is a problem of resource sharing: how can a software system control multiple processes' access to a shared resource, when each process needs exclusive control of that resource while doing its work? The mutual-exclusion solution to this makes the shared resource available only while the process is in a specific code segment called the critical section. It controls access to the shared resource by controlling each mutual execution of that part of its program where the resource would be used. A successful solution to this problem must have at least these two properties: It must implement mutual exclusion: only one process can be in the critical section at a time. It must be free of deadlocks: if processes are trying to enter the critical section, one of them must eventually be able to do so successfully, provided no process stays in the critical section permanently. Deadlock freedom can be expanded to implement one or both of these properties: Lockout-freedom guarantees that any process wishing to enter the critical section will be able to do so eventually. This is distinct from deadlock avoidance, which requires that some waiting process be able to get access to the critical section, but does not require that every process gets a turn. If two processes continually trade a resource between them, a third process could be locked out and experience resource starvation, even though the system is not in deadlock. If a system is free of lockouts, it ensures that every process can get a turn at some point in the future. A k-bounded waiting property gives a more precise commitment than lockout-freedom. Lockout-freedom ensures every process can access the critical section eventually: it gives no guarantee about how long the wait will be. In practice, a process could be overtaken an arbitrary or unbounded number of times by other higher-priority processes before it gets its turn. Under a k-bounded waiting property, each process has a finite maximum wait time. This works by setting a limit to the number of times other processes can cut in line, so that no process can enter the critical section more than k times while another is waiting. Every process's program can be partitioned into four sections, resulting in four states. Program execution cycles through these four states in order: Non-Critical Section Operation is outside the critical section; the process is not using or requesting the shared resource. Trying The process attempts to enter the critical section. Critical Section The process is allowed to access the shared resource in this section. Exit The process leaves the critical section and makes the shared resource available to other processes. If a process wishes to enter the critical section, it must first execute the trying section and wait until it acquires access to the critical section. After the process has executed its critical section and is finished with the shared resources, it needs to execute the exit section to release them for other processes' use. The process then returns to its non-critical section. Enforcing mutual exclusion Hardware solutions On uni-processor systems, the simplest solution to achieve mutual exclusion is to disable interrupts during a process's critical section. This will prevent any interrupt service routines from running (effectively preventing a process from being preempted). Although this solution is effective, it leads to many problems. If a critical section is long, then the system clock will drift every time a critical section is executed because the timer interrupt is no longer serviced, so tracking time is impossible during the critical section. Also, if a process halts during its critical section, control will never be returned to another process, effectively halting the entire system. A more elegant method for achieving mutual exclusion is the busy-wait. Busy-waiting is effective for both uniprocessor and multiprocessor systems. The use of shared memory and an atomic test-and-set instruction provide the mutual exclusion. A process can test-and-set on a location in shared memory, and since the operation is atomic, only one process can set the flag at a time. Any process that is unsuccessful in setting the flag can either go on to do other tasks and try again later, release the processor to another process and try again later, or continue to loop while checking the flag until it is successful in acquiring it. Preemption is still possible, so this method allows the system to continue to function—even if a process halts while holding the lock. Several other atomic operations can be used to provide mutual exclusion of data structures; most notable of these is compare-and-swap (CAS). CAS can be used to achieve wait-free mutual exclusion for any shared data structure by creating a linked list where each node represents the desired operation to be performed. CAS is then used to change the pointers in the linked list during the insertion of a new node. Only one process can be successful in its CAS; all other processes attempting to add a node at the same time will have to try again. Each process can then keep a local copy of the data structure, and upon traversing the linked list, can perform each operation from the list on its local copy. Software solutions In addition to hardware-supported solutions, some software solutions exist that use busy waiting to achieve mutual exclusion. Examples include: Dekker's algorithm Peterson's algorithm Lamport's bakery algorithm Szymański's algorithm Taubenfeld's black-white bakery algorithm Maekawa's algorithm These algorithms do not work if out-of-order execution is used on the platform that executes them. Programmers have to specify strict ordering on the memory operations within a thread. It is often preferable to use synchronization facilities provided by an operating system's multithreading library, which will take advantage of hardware solutions if possible but will use software solutions if no hardware solutions exist. For example, when the operating system's lock library is used and a thread tries to acquire an already acquired lock, the operating system could suspend the thread using a context switch and swap it out with another thread that is ready to be run, or could put that processor into a low power state if there is no other thread that can be run. Therefore, most modern mutual exclusion methods attempt to reduce latency and busy-waits by using queuing and context switches. However, if the time that is spent suspending a thread and then restoring it can be proven to be always more than the time that must be waited for a thread to become ready to run after being blocked in a particular situation, then spinlocks are an acceptable solution (for that situation only). Bound on the mutual exclusion problem One binary test&set register is sufficient to provide the deadlock-free solution to the mutual exclusion problem. But a solution built with a test&set register can possibly lead to the starvation of some processes which become caught in the trying section. In fact, distinct memory states are required to avoid lockout. To avoid unbounded waiting, n distinct memory states are required. Recoverable mutual exclusion Most algorithms for mutual exclusion are designed with the assumption that no failure occurs while a process is running inside the critical section. However, in reality such failures may be commonplace. For example, a sudden loss of power or faulty interconnect might cause a process in a critical section to experience an unrecoverable error or otherwise be unable to continue. If such a failure occurs, conventional, non-failure-tolerant mutual exclusion algorithms may deadlock or otherwise fail key liveness properties. To deal with this problem, several solutions using crash-recovery mechanisms have been proposed. Types of mutual exclusion devices The solutions explained above can be used to build the synchronization primitives below: Locks (mutexes) Readers–writer locks Recursive locks Semaphores Monitors Message passing Tuple space Many forms of mutual exclusion have side-effects. For example, classic semaphores permit deadlocks, in which one process gets a semaphore, another process gets a second semaphore, and then both wait till the other semaphore to be released. Other common side-effects include starvation, in which a process never gets sufficient resources to run to completion; priority inversion, in which a higher-priority thread waits for a lower-priority thread; and high latency, in which response to interrupts is not prompt. Much research is aimed at eliminating the above effects, often with the goal of guaranteeing non-blocking progress. No perfect scheme is known. Blocking system calls used to sleep an entire process. Until such calls became threadsafe, there was no proper mechanism for sleeping a single thread within a process (see polling). See also Atomicity (programming) Concurrency control Dining philosophers problem Exclusive or Mutually exclusive events Reentrant mutex Semaphore Spinlock References Further reading Michel Raynal: Algorithms for Mutual Exclusion, MIT Press, Sunil R. Das, Pradip K. Srimani: Distributed Mutual Exclusion Algorithms, IEEE Computer Society, Thomas W. Christopher, George K. Thiruvathukal: High-Performance Java Platform Computing, Prentice Hall, Gadi Taubenfeld, Synchronization Algorithms and Concurrent Programming, Pearson/Prentice Hall, External links Common threads: POSIX threads explained – The little things called mutexes" by Daniel Robbins Mutual Exclusion with Locks – an Introduction Mutual exclusion variants in OpenMP The Black-White Bakery Algorithm Concurrency control Edsger W. Dijkstra
8302008
https://en.wikipedia.org/wiki/Archive%20bit
Archive bit
The archive bit is a file attribute used by CP/M, Microsoft operating systems, OS/2, and AmigaOS. It is used to indicate whether or not the file has been backed up (archived). Usage In Windows and OS/2, when a file is created or modified, the archive bit is set (i.e. turned on), and when the file has been backed up, the archive bit is cleared (i.e. turned off). Thus, the meaning of the archive bit is "this file has not been archived". In CP/M, the meaning is inverted: creating or modifying a file clears the archive bit, and the bit is set when the file is backed up. Backup software may provide the option to do a full backup while clearing archive bits – that is, to back up all files regardless of their archive bit status, and to clear the archive bit for all files processed by the backup. This allows for the creation of an initial full backup that will be supplemented by incremental backups in the future. Later, the incremental backup task uses the archive bit to distinguish which files have already been backed up, and selects only the new or modified files for backup. The operating system never clears the archive bit unless explicitly told to do so by the user. Even when a user explicitly tells the operating system to copy a file for the purpose of archiving it, the archive bit will still not be cleared. A specific command for clearing or setting the bit must be executed. On the other hand, whenever a file with a clear archive bit is modified (edited), or copied or moved to a new location, the operating system will set the archive bit on the target file. Typically on a file system that has never been formally backed up, nearly all of the files will have their archive bits set, as there are few applications that make use of archive bits without the user's request. In MS-DOS as well as nearly all versions of Windows, status of the archive bit can be viewed or changed with the attrib command-line utility, or by viewing the properties of a file with the Windows shell or Windows Explorer. The archive bit's status can also be viewed or changed with the GetFileAttributes, GetFileAttributesEx, and SetFileAttributes Windows APIs. The archive bit can be set or cleared with the attributes property in VBScript or JScript. The archive bit is a file attribute and not part of the file itself, so the contents of the file remain unchanged when the status of the archive bit changes. Alternatives Relying on the archive bit for backing up files can be unreliable if multiple backup programs are setting and clearing the archive bit on the same volume. Another possibility is to use the timestamp of the last change to the file or directory. This is typically the technique used in UNIX-like operating systems. See also File system File attribute Glossary of backup terms References Computer archives Computer file systems Metadata
35345007
https://en.wikipedia.org/wiki/Email%20hacking
Email hacking
Email hacking is the unauthorized access to, or manipulation of, an account or email correspondence. Overview Email is now a very widely used communication method. If an email account is hacked it can allow the attacker access to the personal, sensitive, or confidential information in the mail storage; as well as allowing them to read new incoming and outgoing email - and to send and receive as the legitimate owner. On some email platforms it may also allow them to setup automated email processing rules. All of these could be very harmful for the legitimate user. Attacks There are a number of ways in which a hacker can illegally gain access to an email account. Virus A virus or other malware can be sent via email, and if executed may be able to capture the user's password and send it to an attacker. Phishing Phishing involves emails that appear to be from legitimate sender but are scams which ask for verification of personal information, such as an account number, a password, or a date of birth. If unsuspecting victims respond, the result may be stolen accounts, financial loss, or identity theft. Prevention measures Email on the internet is sent by the Simple Mail Transfer Protocol (SMTP). While mail can be encrypted between mail servers, this is not typically enforced, but instead Opportunistic TLS is used - where mailservers negotiate for each email connection whether it will be encrypted, and to what standard. Where a mail flow between servers is not encrypted, it could be intercepted by an ISP or government agency and the contents can be read by passive monitoring. For higher security, email administrators can configure servers to require encryption to specified servers or domains. Email spoofing and similar issues which facilitate phishing are addressed by the 'stack' of Sender Policy Framework (SPF), DomainKeys Identified Mail (DKIM) and Domain-based Message Authentication, Reporting and Conformance (DMARC). Setting all these in place is technically challenging, and not visible to the end user, so implementation progress has been slow. A further layer, Authenticated Received Chain (ARC), allows mail flow through intermediate mail servers such as mailing lists or forwarding services to be better handled - a common objection to implementation. Businesses typically have advanced firewalls, anti-virus software and intrusion detection systems (IDS) to prevent or detect improper network access. They may also have security specialists perform an audit on the company and hire a Certified Ethical Hacker to perform a simulated attack or "pen test" in order to find any gaps in security. Although companies may secure its internal networks, vulnerabilities can also occur through home networking. Email may be protected by methods, such as, creating a strong password, encrypting its contents, or using a digital signature. If passwords are leaked or otherwise become known to an attacker, having two-factor authentication enabled may prevent improper access. There are also specialist encrypted email services such as Protonmail or Mailfence. Cases of email hacking Notable cases of email hacks include: Email archives from the Climatic Research Unit were leaked to create the scandal popularly known as Climategate. News of the World journalists hacked email accounts for their stories. UK politician Rowenna Davis had her mail account taken over and held ransom. US politician Sarah Palin has been hacked in order to find embarrassing or incriminating correspondence. As part of the Sony Pictures hack, over 170,000 pieces of email between top executives ended up on Wikileaks. Former United States president, George H.W. Bush had his email hacked. The personal email of political consultant John Podesta was hacked, and contents later published by WikiLeaks. References
1583353
https://en.wikipedia.org/wiki/Pocket%20Tanks
Pocket Tanks
Pocket Tanks is a 1–2 player computer game by Michael P. Welch from Blitwise Productions. Adapted from Michael Welch's earlier Amiga game Scorched Tanks, Pocket Tanks features modified physics, hundreds of weapons ranging from simple explosive shells to homing missiles, and the ability to move the tank. Originally released for Windows and Mac OS X, it was later released also for iOS, Android, and Windows Phone. Gameplay The point of Pocket Tanks, often abbreviated as PTanks or simply as PT, is to use various weapons to attack the other player's tank. Each hit scores a certain number of points, which varies based on the weapon and proximity. At the end of 10 volleys, the player with the highest points wins. The game features a fully destructible environment, which allows player to put themselves on pedestals, in tunnels, or bunkers, allowing for the probability of strategic play. Pocket Tanks supports weapon expansion packs, players can have up to 325 different weapons in total. Network play was made available in the v1.3 release. Expansion packs There are "weapon packs" available which give the player more weapon choices during gameplay. However, these expansion packs are only available to those who own Pocket Tanks Deluxe. For the PC version, there are 30 Expansion Packs available to download on the web page. All packs that include 5 weapons are free, while those with 15 have to be purchased (mobile versions only). There are 325 weapons in total, including 30 from the original game and 295 from the Deluxe Version. Development history Michael P. Welch from Blitwise Productions created Amiga games before he started developing PC games such as Super DX-Ball and Neon Wars. Welch developed Pocket Tanks as remake/sequel of the Amiga game Scorched Tanks and extended the game-play with modified physics, dozens of new weapons, and the ability to move the tank. The game's soundtrack was created by Eliran Ben Ishai as Impulse Tracker Module file. The first version of Pocket Tanks was released in 2001. Version 1.3 was released in November 2007. It contains bug fixes, LAN and IP support. A Collector's Edition of Pocket Tanks was released on November 16, 2007. It contains the first 14 expansion packs and the Deluxe version of Pocket Tanks. The CD and case comes with special art, and it is cheaper to buy the expansion packs through the Collector's Edition, but other than that there are no differences between the Deluxe and Collector's Edition. Version 1.6 is currently the most recent desktop release and it includes more expansion packs to obtain. This version is available for MacOS and the Windows Desktop. On October 26, 2001 the first version of pocket tanks was released after a beta. This version continued to be updated, adding features like network support and more weapons. On June 1, 2008 pocket tanks became available for MacOS. (OS X) On April 15, 2009 pocket tanks was released on iPhone. This version included a new touch interface and 35 weapons to enjoy for free. The iOS version continues to be updated along with other mobile versions that were released later. On December 7, 2012, Pocket Tanks became available on the Android operating system. On January 9, 2013, Blitwise announced the availability of Pocket Tanks on the Windows Store. Users with desktops, laptops, and tablets running Windows 8 or Windows RT can now download the game from the Windows Store. On March 7, 2013, Blitwise announced the availability of "Pocket Tanks" on the Amazon AppStore. Version 2.1 added support for the latest weapon packs and fixes bugs that prevented from being playable on the Surface Pro. Version 2.3 "Brings online functions from a third party Backend as a Service supported framework to a BlitWise maintained AWS hosted server infrastructure." Version 2.5.2 is currently the latest release of the game as of 2020 and is available for Windows Store, iOS, Amazon Appstore, and Android. Pocket Tanks is currently available on Windows, MacOS, iOS, Android, Amazon Appstore, and Windows store. The PC versions of the game lack some of the features of the mobile and Windows store version. The game is still being updated as of 2020. References External links Official site 2001 video games Android (operating system) games Artillery video games IOS games MacOS games Tank simulation video games Video games developed in the United States Windows games Windows Phone games
30261059
https://en.wikipedia.org/wiki/Bill%20Conner
Bill Conner
{{Short description|American businessman} F. William Conner is an American business executive. Education and personal life Conner was born and reared in West Helena, Arkansas. In his early years, he displayed "an interest in energy conversion". In 1981, he received his bachelor's degree from Princeton University in mechanical engineering and later earned an MBA from the Wharton School of the University of Pennsylvania in 1987. Currently, he lives in Dallas, Texas with his wife. Career With a career spanning more than 30 years across high-tech industries, Bill Conner is a corporate turnaround expert and a global leader in cybersecurity, data and infrastructure. From 1981 to 1992, Conner held positions at AT&T including senior engineer, operations manager and strategic planner. Conner held executive positions at Nortel Networks from 1992 to 2001. As the president of Enterprise Data Networks for Nortel, he managed its US$9.1 billion acquisition of Bay Networks. He also served as the first chief marketing officer of Nortel, launching the global marketing campaigns "Come Together" and "What Do You Want the Internet to Be?" Entrust From 2001 to 2013, Conner was the chief executive officer and president of Entrust, a private software company that specializes in identity-based security. Conner was responsible for the creation and delivery of identity-based security software and services. Entrust grew to profitability (albeit short due to its early focus on growth) and reached annual sales of well over $100 million by the time of arrival of Conner to the firm in 2001. Conner shortly thereafter launched a restructuring effort and began to trim workforce annually in an effort to return the company to profitability. He launched a new business model focusing on product portfolios targeted to key verticals and geographies. Spun out of Nortel in 1997, Entrust was a pioneer in public key infrastructure (PKI) technology by providing the first scalable, robust product to offer public key infrastructure for authentication and encryption for web transactions. In April 2002, Entrust's PKI technology served as the foundation for the prototype of what is now the United States Federal Bridge Certification Authority (FBCA), which is a fundamental element of the trust infrastructure that provides the basis for intergovernmental and cross-governmental secure communications. Entrust's PKI is interoperable with all major FBCA vendors. The Entrust Managed PKI service was the first of its kind to deliver certificates utilized by the U.S. Federal Personal Identity Verification (PIV) smartcard, providing authentication, encryption and digital signature to protect U.S. Federal assets from unauthorized access. Under Conner’s leadership, Entrust also played a key role in the introduction of ePassports, including for the United States, where certificates are utilized to protect the privacy of the biometrics while authenticating the citizen. Conner engineered the company's acquisition by private equity firm Thoma Bravo in 2009, and transitioned Entrust into a private firm. In December 2013, it was announced that Datacard Group had entered into an agreement to purchase Entrust for a reported $500 Million. Conner took on a consultant role during the transition. Silent Circle On January 5, 2015, Silent Circle announced that its Board of Directors appointed Conner as Chief Executive Officer and member of the Board of Directors. As CEO of Silent Circle, Conner also joined the Board of Directors of Blackphone. On June 17, 2016, Silent Circle blogged that Conner had resigned as CEO and would remain only as an "advisor," implying he had also resigned his seat on the Board. SonicWall Formerly a division of Dell Software Inc., SonicWall was spun off and acquired by Francisco Partners and Elliott Management in June 2016. On Nov. 1, 2016, SonicWall announced that it had commenced standalone operations, and named Conner its President and CEO. SonicWall states it is the North American market share leader in internet security appliances among small and medium-sized enterprises. In August 2016, SonicWall launched its Capture Advanced Threat Prevention Service, and has analyzed more than 4 million suspicious files for its clients as of Nov. 1, 2016. The cloud-based service uses multiple engines to help prevent attacks, and automates remediation techniques. Conner stated publicly that SonicWall will continue its relationship with Dell, as a reseller partner, and will continue its service areas focused on network security, secure mobile access, email security and encryption, and management and reporting. In a public statement to employees, Conner stated he wants to focus SonicWall on four areas of growth: Next-generation firewall features The user interface, user experience and application program interface of its cloud-based globalization management system (GMS) The SonicOS software operating system and hardware platforms Growing core market segments In September, 2018 at a SonicWall Channel Partner conference, Conner outlined the rapid growth in non-standard port threats, and explained a business reorganization to serve both the SME and enterprise business markets focusing on the enterprise verticals of retail, education and government. To help organizations pivot to remain operational during the COVID-19 pandemic, in April 2020 Conner introduced SonicWall's Boundless Cybersecurity model designed to secure organizations that are increasingly remote, mobile, and less secure, and empower organizations and businesses to close the growing cybersecurity business and skills gaps. Comodo CA (now Sectigo) In October 2017, Francisco Partners acquired Comodo Certification Authority (Comodo CA) and named Bill Conner Chairman of the Board. This marks a transition in ownership from the Comodo organization as Comodo CA, an industry leader in SSL security with more than 200,000 customers across 150 countries, focuses on the significant market opportunity associated with SSL certificates and the massive growth in online and IoT traffic. In November 2018, Comodo CA rebranded itself as Sectigo, to avoid confusion with its former parent entity Comodo Group. As a provider of SSL and other types of digital certificates, Sectico is also heavily invested in PKI. With the new name came a large investment in X.509 certificate management, including the discovery and reporting of certificates issued via PKI but also the automated installation of certificates, which helps saves labor cost while eliminating mistakes made by mis-configured and expired certificates. In 2019, Sectigo was the first commercial CA to delivered automated installation of both Public and Private SSL certificates into web servers and load balancers on both sides of the firewall using ACME (RFC8555), in collaboration with the Electronic Frontier Foundation (EFF). Sectigo also released industry-first PKI DevOps integrations for management and container orchestration platforms, including Docker, Kubernetes, Ansible and Terraform. In 2020, GI Partners announced a definitive agreement to acquire Sectigo, with the intention of supporting the acceleration of Sectigo’s security innovations and growth and fostering worldwide expansion. Cybersecurity and infrastructure global trends Conner has been active in various public-private partnerships on cybersecurity, cybertheft and infrastructure security. He helped unveil the INTERPOL Global Smart eID Card and addressed the United Nations on global challenges in cybercrime in June 2010 He co-chaired the Corporate Governance Task Force of the U.S. Department of Homeland Security National Cybersecurity Partnership He has been a member of the Business Software Alliance and the Cyber Security Industry Alliance (now a branch of TechAmerica technology trade association) He created and co-chaired the Information Security Governance Task Force of the Business Software Alliance He is a member of the Enterprise Software Roundtable In a September 2018 article on Cybersecurity Trends, Conner describes innovative next-generation security techniques in the cloud and the deep web. The article references RTDMI (Real-Time Deep Memory Inspection), a new patent-pending technology created by SonicWall. In February 2018, Former Secretary of the Department of Homeland Security Michael Chertoff and Bill Conner collaborated on an opinion article written for The Hill. In it, they detail some of the recent actions taken by the U.S. Congress and the SEC regarding cyber accountability and transparency. In 2017, Conner became a Forbes Community Voice Contributor, as a member of the Forbes Technology Council. His articles have featured topics such as cybersecurity and policy trends, cyber threat intelligence, and how human predictability factors in cyberattacks. He has also penned articles and contributed to reports on email-borne cyberattacks and ransomware. In February 2012, Conner took part in a United States House of Representatives committee on Energy and Commerce hearing to discuss cybersecurity threats facing the US and the role of public-private partnerships play in defense. Conner has recently spent time discussing the effects of spear phishing schemes on weak authentication, notably the rash of attacks on media outlets, and the risks of insider threats. Recent Publications "Webinar (Italy): 2021 Cyber Threat Landscape Roundtable," Fondazione CRUI, March 26, 2021. “Cybersecurity for the post-COVID new normal of work,” Harvard Business School – Managing the Future of Work (audio podcast), August 25, 2020 “Veiled Threats: The Growing Cyberattack Vectors Few People Talk About,” "Forbes," Bill Conner (author), Dec. 6, 2019 “Back-To-School Lists Should Now Include Ransomware,” "Forbes," Bill Conner (author), Sept. 24, 2019 “Bill Conner: You Cannot Have Privacy Without Security,” Cyber Security Interviews (audio podcast), Sept. 16, 2019 “The CyberWire Daily Podcast: Episode 926,” The Cyberwire Daily Podcast (audio podcast), Sept. 12, 2019 "Side-Channel Attacks: Cyber Warfare’s New Battleground,” “Security Boulevard,” Bill Conner (author), Aug. 22, 2019 "The Independent Show,” Keynote Speaker, ACA Connects Event, July 31, 2019 "AI, Threat Intelligence and The Cyber Arms Race: SonicWall CEO Bill Conner Joins The Chertoff Group Security Series," "The Chertoff Group," (video of panel discussion), June 19, 2019 "In focus: SonicWall CEO Bill Conner," SC Magazine, May 7, 2019 "Dramatic rise in fraudulent PDF files in 2019: SonicWall study," "CISCO Mag," Apr. 22, 2019 “Old-school cruel: Dodgy PDF email attachments enjoying a renaissance,” “The Register,” John Oates (author), Apr. 19, 2019 “Report: Weaponized PDFs on the Rise,” “Nextgov,” Brandi Vincent (author), Apr. 19, 2019 “Insights & Intelligence Podcast: Inside the Tactical Advances between Cybercriminals and Security Industry,” “The Chertoff Group,” (audio podcast), Apr. 15, 2019 “SonicWall report paints sobering picture of cyberthreat trends,” “SiliconAngle,” Paul Gillin (author), Mar, 26, 2019 "Slate Of New Product Offerings Marks Rebirth Of SonicWall," "Forbes," Tony Bradley (author), Feb. 27, 2019 "Who’s taking malware seriously? SonicWall’s CEO has the answers," "Information Age," Nick Ismael (author), Feb. 7, 2019 "SMBs Need Layered Security to Defend Their Businesses," "Forbes," Bill Conner (author), Feb. 4, 2019 "CEO Outlook: 5 Questions on 2019," "CRN," Bill Conner (author), January 2019 "New Law Aids SMBs in Combating Cybersecurity Risks," "ChannelPro Network," Martin Sinderman (author), Dec. 6, 2018 "CEO Spotlight: Bill Conner, CEO, SonicWall," "KRLD Radio," Bill Conner (author), Nov. 30, 2018 “Opinion: Two cybersecurity policies, one clear new objective,” The Hill, Bill Conner (author), Nov. 20, 2018 “Congress Passes Bill Creating Cybersecurity Agency at DHS,” Security Week, Eduard Kovacs (author), Nov. 15, 2018 “13 Tech Experts Share What Facebook Should Do Post-Data Breach,” Forbes, Bill Connor (co-author), Nov. 13, 2018 “Privacy Problems for FANG Companies Might Beget M&A Action in Cybersecurity,” The Street, Kevin Curran (author), Oct. 25, 2018 “Facebook hack: How attackers stole the personal data of almost 30 million users,” ITPro (UK), Keumars Afifi-Sabet (author), Oct. 22, 2018 “Chart of the Day: Google Plus Never Got Off the Ground,” Real Money, Kevin Curran (author), Oct. 9, 2018 “Hackers compromise Facebook users’ security,” The Signal (US), James Wright (author), Oct. 8, 2018 “ ‘A leader has to be passionate’, ” Authority Magazine, Oct. 7, 2018 “ ‘Unprecedented’ Russian cyberattacks announcement is just the start – and businesses should be prepared,” Verdict, Lucy Ingham (author) Oct. 4, 2018 “UK and allies accuse Russia of cyber attack campaign,” Computer Weekly, Warwick Ashford (author) Oct. 4, 2018 “The UK, US and Netherlands accuse Russia of cyber attacks,” ITPro (UK), Bobby Hellard (author) Oct. 5, 2018 “Betabot infostealer infections spreading in the wild, researchers say,” SC Magazine, Rene Millman (author) Oct. 3, 2018 “Facebook Hack: People's Accounts Appear for Sale On Dark Web,” The Independent, Anthony Cuthbertson (author) Oct. 2, 2018 “Facebook Hack May Result in GDPR Penalty,” Silicon (UK), Tom Jowitt (author) Oct. 2, 2018 "Bill Conner: A Vanguard of Transformation Technology Shifts," Insights Success, The 10 Most Admired CEO's to Watch, August, 2018 "Protecting SMBs: The Key to Shrinking Nation’s Cyberattack Surface," Money, Inc., Bill Conner (author) Sept. 19 2018 In this article, Conner describes the impact of the new National Institute of Standards and Technology (NIST) Small Business Cybersecurity Act. "Social Media Security: A Costly, but Worthy Cause for Shareholders," Real Money, Kevin Curran (author) Sept. 20, 2018 "How machine learning allowed one company to detect Meltdown and Spectre before Intel went public," TechRepublic, Jason Hiner (author) Apr. 20, 2018 "The Value of Enterprise Privacy," The Guardian, Jul. 23, 2015 "Viewpoint: U.K. Breach a Precautionary Tale for U.S.," American Banker, December 2007 "Cybersecurity: Business Has to Enlist in This War" (co-author), guest opinion column, New York Times, Nov. 17, 2003 "Vulnerability on the Cyber Front," co-author with Ret. Gen. Wesley Clark, Washington Times, Aug. 19, 2002 Awards The Top 25 IT Innovators of 2021, CRN, August 2, 2021. SonicWall Attains 5-Star Rating in 2021 CRN Partner Program Guide, CRN, March 30, 2021. 2021 Cybersecurity Global Excellence Awards: Grand Trophy, The Globee Awards, March 3, 2021. Top 25 Enterprise IT Innovators of 2019, CRN. 11 Coolest Endpoint Security CEOs of 2019, Solutions Review, July, 2019. 10 Most Admired CEO's to Watch, Insights Success, August, 2018. Bill Conner, President and CEO of SonicWall, Chairman of the Board of Directors of Comodo CA, named by Info Security Product Guide as 2018 CEO of the Year (500-2,499 Employees): Gold Winner. CEO World Awards 2018 – CEO of the Year (500-2,499 Employees): Silver Winner. Bill Conner honored for driving SonicWall to surpass financial objectives for six consecutive quarters since becoming independent from Dell. Listed by CRN as One of the Most Influential Executives of 2018 (cited as #21) 2017 SC Media 2017 Reboot Leadership Award to Bill Conner One of the 25 Most Influential Executives Of 2017 according to CRN CRN Names Bill Conner 2017 Top Midmarket Vendor Executive 2015 Global New Product Innovation Award to Silent Circle: Frost & Sullivan 2015: Global award given by Frost & Sullivan based on customer impact and new product attributes. Top 10 Most Innovative Companies of 2015 Fast Company: Silent Circle: Consumer Electronics focused review of Global companies leading in technology and security. Best of Mobile World Congress 2015: PCMag.com 2015: For Silent Circle BlackPhone 2: Top smartphone or tablet recognition from national magazine on technology products. 2014 ICON Technology of the Year Award: for BlackPhone, Silent Circle and Geeksphone: ICON Magazine U.S. Federal 100 Award: Federal Computer Week March 2004: National award from Federal Computer Week recognizing the top 100 individuals and organizations for impact on U.S. Government technology and services. Tech Titans Corporate CEO of the Year: Metroplex Technology Business Council September 2003: Given by Metroplex Technology Business Council, recognizing technology leadership and raising scholarship funds for technology students. References Living people American technology chief executives Businesspeople in software Wharton School of the University of Pennsylvania alumni Princeton University School of Engineering and Applied Science alumni Businesspeople from Dallas People from West Helena, Arkansas Nortel employees American mechanical engineers AT&T people Chief marketing officers Year of birth missing (living people) Place of birth missing (living people)
729762
https://en.wikipedia.org/wiki/RSBAC
RSBAC
Rule-set-based access control (RSBAC) is an open source access control framework for current Linux kernels, which has been in stable production use since January 2000 (version 1.0.9a). Features Free open source GNU General Public License (GPL) Linux kernel security extension Independent of governments and big companies Several well-known and new security models, e.g. mandatory access control (MAC), access control list (ACL), and role compatibility (RC) On-access virus scanning with Dazuko interface Detailed control over individual user and program network accesses Fully access controlled kernel level user management Any combination of security models possible Easily extensible: write your own model for runtime registration Support for latest kernels Stable for production use Easily portable to other operating systems The RSBAC system architecture has been derived and extended from the Generalized Framework for Access Control (GFAC) by Marshall Abrams and Leonard La Padula. RSBAC means "ruleset based access control" and is also a role-based access control (RBAC) solution. The two acronyms can cause confusion. In his essay "Rule Set Modeling of a Trusted Computer System", Leonard LaPadula describes how the Generalized Framework for Access Control (GFAC) approach could be implemented in the UNIX System V operating system. He introduced the clear separation between Access Enforcement Facility (AEF), Access Decision Facility (ADF) with Access Control Rules (ACR), and Access Control Information (ACI). The AEF as part of the system call function calls the ADF, which uses ACI and the rules to return a decision and a set of new ACI attribute values. The decision is then enforced by the AEF, which also sets the new attribute values and, in case of allowed access, provides object access to the subject. This structure requires all security relevant system calls to be extended by AEF interception, and it needs a well-defined interface between AEF and ADF. For better modeling, a set of request types was used in which all system call functionalities were to be expressed. The general structure of the GFAC has also been included in the ISO standard 10181-3 Security frameworks for open systems: Access control framework and into The Open Group standard Authorization (AZN) API. The first RSBAC prototype followed La Padula's suggestions and implemented some access control policies briefly described there, namely mandatory access control (MAC), functional control (FC) and Security Information Modification (SIM), as well as the Privacy Model by Simone Fischer-Hübner. Many aspects of the system have changed a lot since then, e.g. the current framework supports more object types, includes generic list management and network access control, contains several additional security models, and supports runtime registration of decision modules and system calls for their administration. RSBAC and other solutions RSBAC is very close to Security-Enhanced Linux (SELinux), as they share a lot more in their design than other access controls such as AppArmor. However, RSBAC brings its own hooking code instead of relying on the Linux Security Module (LSM). Due to this, RSBAC is technically a replacement for LSM itself, and implement modules that are similar to SELinux, but with additional functionality. The RSBAC framework incorporates complete object status and has a full knowledge of the kernel state when making decisions, making it more flexible and reliable. However, this comes at the cost of slightly higher overhead in the framework itself. Although SELinux- and RSBAC-enabled systems have similar impact on performance, LSM impact alone is negligible compared to the RSBAC framework alone. For this reason, LSM has been selected as default and unique security-hooking mechanism in the Linux kernel, RSBAC coming as a separate patch only. History RSBAC was the first Linux role-based access control (RBAC) and mandatory access control (MAC) patch. See also External links RSBAC homepage Leonard LaPadula's essay on GFAC Gentoo Wiki Access Control Comparison Table Operating system security Linux security software Linux kernel features Access control
2657086
https://en.wikipedia.org/wiki/Millie%20Small
Millie Small
Millicent Dolly May Small CD (6 October 1947 – 5 May 2020) was a Jamaican singer and songwriter, best known for her 1964 recording of "My Boy Lollipop", which reached number two in both the UK Singles Chart and the US Billboard Hot 100. On her UK records, she was usually credited mononymously as Millie. She was the Caribbean's first international recording star, and its most successful female performer. Early life and career Millie Small was born on 6 October 1947 in Clarendon, Jamaica, the daughter of a sugar plantation overseer. She was one of 13 siblings, with seven brothers and five sisters. Like many Jamaican singers of the era, her career began by winning the Vere Johns Opportunity Hour talent contest, which she won at the age of twelve. Wishing to pursue a career as a singer, she moved to live with relatives in Love Lane in Kingston. She auditioned for Studio One record producer Coxsone Dodd, who was struck by the similarity of her voice to that of Shirley Goodman of the American duo Shirley and Lee. He paired her with singer Owen Gray, and they made several records together, including "Sugar Plum", which became a local hit. When Gray resumed his solo career, Small began recording with another singer, Samuel Augustus "Roy" Panton. Working with producer Roy Robinson, the duo of Roy & Millie had a run of local hits, including "We'll Meet". They had further successes working with Dodd, as well with producer Lindon Pottinger, including the local hit "Marie" in 1963; and then with Prince Buster. Her popularity brought her to the attention of Anglo-Jamaican entrepreneur Chris Blackwell, who was convinced of her wider international potential, and became her manager and legal guardian. In late 1963 he took her to Forest Hill, London, where she was given intensive training in dancing and diction. International success Her first recording in London, "Don't You Know", made little impact when released by Fontana Records in late 1963, but for her next recording Blackwell recruited guitarist and arranger Ernest Ranglin to oversee the session. Ranglin and his musicians adopted the newly-popular ska style, and his rearrangement of "My Boy Lollipop", a song originally released in the US by teenager Barbie Gaye in late 1956, became immediately successful. Released in March 1964, Small's version (on which she was credited simply as "Millie") was a massive hit. She appeared on British TV shows including Top of the Pops, and the single reached number two in the UK Singles Chart, in the US Billboard Hot 100, and in Canada. It also topped the chart in Australia. Initially it sold over 600,000 copies in the United Kingdom. Including singles sales, album usage, and compilation inclusions, the song has since sold more than seven million copies worldwide and made Small an international star at the age of 17. "My Boy Lollipop" was doubly significant in British pop history. It was the first major hit for Island Records (although it was actually released on the Fontana label because Chris Blackwell, Island's owner, did not want to overextend its then-meagre resources; in the US, the record appeared on the Smash Records subsidiary of Mercury Records). Small was the first artist to have a hit that was recorded in the bluebeat style, a music genre that was a direct ancestor of reggae. She was billed as "The Blue Beat Girl" on the single's label in the US. She toured in Britain and appeared frequently on British television, before collapsing from exhaustion and food poisoning; she was also involved in a traffic accident. Although her next single, "Sweet William", was less successful, reaching number 30 in the UK, number 40 in the US, and number 22 in Canada, she had become an international celebrity. She was given a gold disc in New York, and was driven in an open-top car on her return to Kingston where she performed in several major shows, on one occasion topping the bill over Otis Redding, Patti LaBelle, and Inez and Charlie Foxx. She also performed in a Ska Spectacular show at the 1964 New York World's Fair. Her first album, More Millie, contained a varied selection of songs arranged by Ranglin. In the US it was issued as My Boy Lollipop, with a slightly different selection of tracks. Her early Jamaican recordings were also reissued to take advantage of her popularity. She appeared on the 1964 Beatles TV special Around the Beatles. On 28 December 1964 she also appeared in ITV's Play of the Week episode "The Rise and Fall of Nellie Brown", playing the role of Selina Brown. The play features a number of songs composed by Dolores Claman and performed by Small. In early 1965, she featured in a Ready, Steady, Go! special, Millie in Jamaica, and soon afterwards she embarked on a world tour, with concerts in New Zealand, Australia, Hong Kong, Singapore, Japan, the US, Brazil and Argentina. On 6 March 1965, Small appeared on the Australian television programme Bandstand. This was as part of a concert at the Sidney Myer Music Bowl in Kings Domain, Melbourne, part of the Moomba Festival. She performed "My Boy Lollipop", "What Am I Living For", and "See You Later, Alligator". She recorded several duets with Jackie Edwards in 1965, as well as solo material, but sales were disappointing. She also released her second LP record, Millie Sings Fats Domino, but it failed to reach the charts. Her eighth single in Britain, a version of Wynonie Harris' "Bloodshot Eyes", was her last UK chart success, reaching number 48 in late 1965, but she continued to tour successfully in Australia and Africa. After returning to Britain she made further recordings with Jackie Edwards, including the album Pledging My Love, and also appeared on the compilation album Ska at the Jamaica Playboy Club, singing on one track with the then-unknown Jimmy Cliff. However, her popularity in Britain appeared to dwindle as she spent more time touring abroad. In 1968, after two albums of The Best of Jackie & Millie, her recording contracts with Island and Fontana ended. The emergence of reggae in the United Kingdom in 1969 prompted a return to recording for Small, with the single "My Love and I", on which she was backed by the band Symarip. She then recorded for the Trojan label, her first single combining a version of Nick Drake's "Mayfair" with her own song, the politically-inspired and defiant "Enoch Power", which faced a radio ban but re-established her profile among the British Caribbean community. However, after a short period with President Records, she ended her recording career soon afterwards. Small continued to tour and perform in Jamaica, but in 1971 decided to move to Singapore to live. She returned to Britain in 1973, to coincide with the release of another compilation album, Lollipop Reggae. Thereafter, she largely stayed out of the public eye, even when "My Boy Lollipop" was reissued and re-charted in the UK in 1987 at no. 46. Later life In 1987, during a rare interview with Thames News, it was revealed that Small was destitute and had taken to living in a hostel with her toddler daughter. In November 1987, she made a rare public appearance in Jamaica to receive the Medal of Appreciation from Prime Minister Edward Seaga. In 2006, she was said to be making new recordings after some years spent writing, painting, and raising her daughter. On 6 August 2011, the 49th anniversary of Jamaica's independence, the Governor-General made Small a Commander in the Order of Distinction for her contribution to the Jamaican music industry. The award was accepted on her behalf by Seaga. In July 2012, she again stated that she had been recording again and planned to perform in Jamaica for the first time in over 40 years. After nearly forty years away from the limelight and refusing most interview requests, Small granted U.S. journalist Tom Graves the first ever in-depth interview in the August 2016 edition of Goldmine. Previous interviews were typically short and based on press releases. In the interview she discusses fully her early career and the full impact of "My Boy Lollipop". She also insisted, against the denials of Rod Stewart, that it was Stewart who played harmonica on "My Boy Lollipop". She said she remembered the sessions well and recalls Stewart being asked to play. Small also said in 2016 that she had not received any royalties for the single. Personal life She had a brief relationship with Peter Asher of the 1960s duo Peter & Gordon. In her August 2016 interview with U.S. journalist Tom Graves she said the relationship had been platonic. She lived in Singapore from 1971 to 1973 before returning to the United Kingdom, where she lived for the remainder of her life. She had a daughter, Jaelee, born in 1984, who studied art and the music industry and is a singer-songwriter. Death Small died on 5 May 2020 in London, from a stroke, aged 72. News of her death was first announced to the Jamaica Observer by Chris Blackwell, who last met Small some 12 years before her death. He remembered her as "a sweet person" with a "great sense of humour". Blackwell also credited her for popularising ska on an international level from 1964 as "it was her first hit record". Discography Albums My Boy Lollipop (1964, UK: Fontana, US: Smash) Time Will Tell (1970/2004, Trojan) My Boy Lollipop and 31 Other Songs (1994, DE: Combo Records) Singles Source: See also Caribbean music in the United Kingdom List of performers on Top of the Pops List of reggae musicians List of ska musicians List of stage names Music of Jamaica References External links 1964 biographical article from Teenville magazine Article on her 60th birthday in Jamaica Gleaner Discography at MusicMatch website Millie Small at Webstarts The Millie Small TV Show on Finnish Television, 1964 1947 births 2020 deaths 20th-century Jamaican women singers Afro-Jamaican Island Records artists Smash Records artists People from Clarendon Parish, Jamaica Trojan Records artists Commanders of the Order of Distinction Ska musicians Jamaican reggae singers Deaths from cerebral infarction Jamaican expatriates in the United Kingdom Fontana Records artists Decca Records artists
5778780
https://en.wikipedia.org/wiki/Dynamic%20Packet%20Transport
Dynamic Packet Transport
Dynamic Packet Transport (DPT) is a Cisco transport protocol designed for use in optical fiber ring networks. In overview, it is quite similar to POS and DTM. It was one of the major influences on the Resilient Packet Ring/802.17 standard. Protocol design DPT is implemented as two counter-rotating rings. This means the network is composed of two completely separate rings of fiber that are both able to transmit data concurrently. This design provides for redundancy in case of a fiber cut or link failure, and increased throughput in common situations. DPT as opposed to POS or normal SONET/SDH is able to use both rings at the same time whereas POS only uses one ring under normal circumstances but switches to the second upon failure of the first. Cisco claims that DPT can run with double the bit-rate of POS due to this characteristic. DPT is not a PPP whereas POS is, this means that traffic between two nodes of a DPT ring does not affect intermediate nodes. With the introduction of DPT came the introduction of another Cisco developed MAC layer protocol, Spatial Reuse Protocol or SRP. The use of SRP in conjunction with DPT makes it possible for DPT to communicate with the physical layer. Types of data in DPT networks As with most other lower layer protocols, there are methods for communicating not only application data between the nodes of a DPT network. It is necessary for the nodes to be able to communicate control data between each other in case of a fiber cut or link failure so the nodes can forward traffic on the appropriate interfaces and maintain network connectivity. Both control packets, and data packets are transmitted on both rings in order to maintain connectivity and full bandwidth utilization in normal situations; but once a failure occurs, the control data will notify the applicable routers of the failure and all the routers will switch to using only their active interfaces for data and control packets. DPT packet structure The structure of a DPT Packet is quite similar to that of Ethernet. It contains a source and destination MAC address (both 48-bits long), a protocol type identifier (used for identifying the upper layer protocol contained in the payload), and a frame check sequence used to validate the data. DPT topologies Both DPT and SRP are independent of their physical layers. This means that the DPT protocol can operate above several physical mediums such as SONET/SDH, Gigabit Ethernet, and others. As aforementioned, DPT is composed of two rings for fault tolerance and increased throughput. The method for switching between these two rings in the event of a failure is called Intelligent Protection Switching, or IPS. This ensures that a fiber cut or link failure (layer 1 error) will be rectified and IP traffic will be resumed within 50 ms. DPT also contains a "plug and play" feature which dynamically fetches the MAC addresses of neighboring devices which provides for very simple configuration with little to no setup prior to functional data transfer. Data transmission Packets (information technology)
1609031
https://en.wikipedia.org/wiki/IDN%20homograph%20attack
IDN homograph attack
The internationalized domain name (IDN) homograph attack is a way a malicious party may deceive computer users about what remote system they are communicating with, by exploiting the fact that many different characters look alike (i.e., they are homographs, hence the term for the attack, although technically homoglyph is the more accurate term for different characters that look alike). For example, a regular user of example.com may be lured to click a link where the Latin character "a" is replaced with the Cyrillic character "а". This kind of spoofing attack is also known as script spoofing. Unicode incorporates numerous writing systems, and, for a number of reasons, similar-looking characters such as Greek Ο, Latin O, and Cyrillic О were not assigned the same code. Their incorrect or malicious usage is a possibility for security attacks. The registration of homographic domain names is akin to typosquatting, in that both forms of attacks use a similar-looking name to a more established domain to fool a user. The major difference is that in typosquatting the perpetrator attracts victims by relying on natural typographical errors commonly made when manually entering a URL, while in homograph spoofing the perpetrator deceives the victims by presenting visually indistinguishable hyperlinks. Indeed, it would be a rare accident for a web user to type, for example, a Cyrillic letter within an otherwise English word such as "citibаnk". There are cases in which a registration can be both typosquatting and homograph spoofing; the pairs of l/I, i/j, and 0/O are all both close together on keyboards and, depending on the typeface, may be difficult or impossible to distinguish. History An early nuisance of this kind, pre-dating the Internet and even text terminals, was the confusion between "l" (lowercase letter "L") / "1" (the number "one") and "O" (capital letter for vowel "o") / "0" (the number "zero"). Some typewriters in the pre-computer era even combined the L and the one; users had to type a lowercase L when the number one was needed. The zero/o confusion gave rise to the tradition of crossing zeros, so that a computer operator would type them correctly. Unicode may contribute to this greatly with its combining characters, accents, several types of hyphen, etc., often due to inadequate rendering support, especially with smaller font sizes and the wide variety of fonts. Even earlier, handwriting provided rich opportunities for confusion. A notable example is the etymology of the word "zenith". The translation from the Arabic "samt" included the scribe's confusing of "m" into "ni". This was common in medieval blackletter, which did not connect the vertical columns on the letters i, m, n, or u, making them difficult to distinguish when several were in a row. The latter, as well as "rn"/"m"/"rri" ("RN"/"M"/"RRI") confusion, is still possible for a human eye even with modern advanced computer technology. Intentional look-alike character substitution with different alphabets has also been known in various contexts. For example, Faux Cyrillic has been used as an amusement or attention-grabber and "Volapuk encoding", in which Cyrillic script is represented by similar Latin characters, was used in early days of the Internet as a way to overcome the lack of support for the Cyrillic alphabet. Another example is that vehicle registration plates can have both Cyrillic (for domestic usage in Cyrillic script countries) and Latin (for international driving) with the same letters. Registration plates that are issued in Greece are limited to using letters of the Greek alphabet that have homoglyphs in the Latin alphabet, as European Union regulations require the use of Latin letters. Homographs in ASCII ASCII has several characters or pairs of characters that look alike and are known as homographs (or homoglyphs). Spoofing attacks based on these similarities are known as homograph spoofing attacks. For example, 0 (the number) and O (the letter), "l" lowercase L, and "I" uppercase "i". In a typical example of a hypothetical attack, someone could register a domain name that appears almost identical to an existing domain but goes somewhere else. For example, the domain "rnicrosoft.com" begins with "r" and "n", not "m". Other examples are G00GLE.COM which looks much like GOOGLE.COM in some fonts. Using a mix of uppercase and lowercase characters, googIe.com (capital i, not small L) looks much like google.com in some fonts. PayPal was a target of a phishing scam exploiting this, using the domain PayPaI.com. In certain narrow-spaced fonts such as Tahoma (the default in the address bar in Windows XP), placing a c in front of a j, l or i will produce homoglyphs such as cl cj ci (d g a). Homographs in internationalized domain names In multilingual computer systems, different logical characters may have identical appearances. For example, Unicode character U+0430, Cyrillic small letter a ("а"), can look identical to Unicode character U+0061, Latin small letter a, ("a") which is the lowercase "a" used in English. Hence wikipediа.org (xn--wikipedi-86g.org; the Cyrillic version) instead of wikipedia.org (the Latin version). The problem arises from the different treatment of the characters in the user's mind and the computer's programming. From the viewpoint of the user, a Cyrillic "а" within a Latin string is a Latin "a"; there is no difference in the glyphs for these characters in most fonts. However, the computer treats them differently when processing the character string as an identifier. Thus, the user's assumption of a one-to-one correspondence between the visual appearance of a name and the named entity breaks down. Internationalized domain names provide a backward-compatible way for domain names to use the full Unicode character set, and this standard is already widely supported. However this system expanded the character repertoire from a few dozen characters in a single alphabet to many thousands of characters in many scripts; this greatly increased the scope for homograph attacks. This opens a rich vein of opportunities for phishing and other varieties of fraud. An attacker could register a domain name that looks just like that of a legitimate website, but in which some of the letters have been replaced by homographs in another alphabet. The attacker could then send e-mail messages purporting to come from the original site, but directing people to the bogus site. The spoof site could then record information such as passwords or account details, while passing traffic through to the real site. The victims may never notice the difference, until suspicious or criminal activity occurs with their accounts. In December 2001 Evgeniy Gabrilovich and Alex Gontmakher, both from Technion, Israel, published a paper titled "The Homograph Attack", which described an attack that used Unicode URLs to spoof a website URL. To prove the feasibility of this kind of attack, the researchers successfully registered a variant of the domain name microsoft.com which incorporated Cyrillic characters. Problems of this kind were anticipated before IDN was introduced, and guidelines were issued to registries to try to avoid or reduce the problem. For example, it was advised that registries only accept characters from the Latin alphabet and that of their own country, not all of Unicode characters, but this advice was neglected by major TLDs. On February 7, 2005, Slashdot reported that this exploit was disclosed by 3ric Johanson at the hacker conference Shmoocon. Web browsers supporting IDNA appeared to direct the URL http://www.pаypal.com/, in which the first a character is replaced by a Cyrillic а, to the site of the well known payment site PayPal, but actually led to a spoofed web site with different content. Popular browsers continued to have problems properly displaying international domain names through April 2017. The following alphabets have characters that can be used for spoofing attacks (please note, these are only the most obvious and common, given artistic license and how much risk the spoofer will take of getting caught; the possibilities are far more numerous than can be listed here): Cyrillic Cyrillic is, by far, the most commonly used alphabet for homoglyphs, largely because it contains 11 lowercase glyphs that are identical or nearly identical to Latin counterparts. The Cyrillic letters а, с, е, о, р, х and у have optical counterparts in the basic Latin alphabet and look close or identical to a, c, e, o, p, x and y. Cyrillic З, Ч and б resemble the numerals 3, 4 and 6. Italic type generates more homoglyphs: дтпи or дтпи (дтпи in standard type), resembling dmnu (in some fonts д can be used, since its italic form resembles a lowercase g; however, in most mainstream fonts, д instead resembles a partial differential sign, ∂). If capital letters are counted, АВСЕНІЈКМОРЅТХ can substitute ABCEHIJKMOPSTX, in addition to the capitals for the lowercase Cyrillic homoglyphs. Cyrillic non-Russian problematic letters are і and i, ј and j, ԛ and q, ѕ and s, ԝ and w, Ү and Y, while Ғ and F, Ԍ and G bear some resemblance to each other. Cyrillic ӓёїӧ can also be used if an IDN itself is being spoofed, to fake äëïö. While Komi De (ԁ), shha (һ), palochka (Ӏ) and izhitsa (ѵ) bear strong resemblance to Latin d, h, l and v, these letters are either rare or archaic and are not widely supported in most standard fonts (they are not included in the WGL-4). Attempting to use them could cause a ransom note effect. Greek From the Greek alphabet, only omicron ο and sometimes nu ν appear identical to a Latin alphabet letter in the lowercase used for URLs. Fonts that are in italic type will feature Greek alpha α looking like a Latin a. This list increases if close matches are also allowed (such as Greek εικηρτυωχγ for eiknptuwxy). Using capital letters, the list expands greatly. Greek ΑΒΕΗΙΚΜΝΟΡΤΧΥΖ looks identical to Latin ABEHIKMNOPTXYZ. Greek ΑΓΒΕΗΚΜΟΠΡΤΦΧ looks similar to Cyrillic АГВЕНКМОПРТФХ (as do Cyrillic Л (Л) and Greek Λ in certain geometric sans-serif fonts), Greek letters κ and о look similar to Cyrillic к and о. Besides this Greek τ, φ can be similar to Cyrillic т, ф in some fonts, Greek δ resembles Cyrillic б in the Serbian alphabet, and the Cyrillic а also italicizes the same as its Latin counterpart, making it possible to substitute it for alpha or vice versa. The lunate form of sigma, Ϲϲ, resembles both Latin Cc and Cyrillic Сс. If an IDN itself is being spoofed, Greek beta β can be a substitute for German eszett ß in some fonts (and in fact, code page 437 treats them as equivalent), as can Greek end-of-word-variant sigma ς for ç; accented Greek substitutes όίά can usually be used for óíá in many fonts, with the last of these (alpha) again only resembling a in italic type. Armenian The Armenian alphabet can also contribute critical characters: several Armenian characters like օ, ո, ս, as well capital Տ and Լ are often completely identical to Latin characters in modern fonts. Symbols like ա can resemble Cyrillic ш. Beside that, there are symbols which look alike. ցհոօզս which look like ghnoqu, յ which resembles j (albeit dotless), and ք, which can either resemble p or f depending on the font. However, the use of Armenian is problematic. Not all standard fonts feature the Armenian glyphs (whereas the Greek and Cyrillic scripts are in most standard fonts). Because of this, Windows prior to Windows 7 rendered Armenian in a distinct font, Sylfaen, which supports Armenian, and the mixing of Armenian with Latin would appear obviously different if using a font other than Sylfaen or a Unicode typeface. (This is known as a ransom note effect.) The current version of Tahoma, used in Windows 7, supports Armenian (previous versions did not). Furthermore, this font differentiates Latin g from Armenian ց. Two letters in Armenian (Ձշ) also can resemble the number 2, Յ resembles 3, while another (վ) sometimes resembles the number 4. Hebrew Hebrew spoofing is generally rare. Only three letters from that alphabet can reliably be used: samekh (ס), which sometimes resembles o, vav with diacritic (וֹ), which resembles an i, and heth (ח), which resembles the letter n. Less accurate approximants for some other alphanumerics can also be found, but these are usually only accurate enough to use for the purposes of foreign branding and not for substitution. Furthermore, the Hebrew alphabet is written from right to left and trying to mix it with left-to-right glyphs may cause problems. Thai Though the Thai script has historically had a distinct look with numerous loops and small flourishes, modern Thai typography, beginning with Manoptica in 1973 and continuing through IBM Plex in the modern era, has increasingly adopted a simplified style in which Thai characters are represented with glyphs strongly representing Latin letters. ค (A), ท (n), น (u), บ (U), ป (J), พ (W), ร (S), and ล (a) are among the Thai glyphs that can closely resemble Latin. Chinese The Chinese language can be problematic for homographs as many characters exist as both traditional (regular script) and simplified Chinese characters. In the .org domain, registering one variant renders the other unavailable to anyone; in .biz a single Chinese-language IDN registration delivers both variants as active domains (which must have the same domain name server and the same registrant). .hk (.香港) also adopts this policy. Other scripts Other Unicode scripts in which homographs can be found include Number Forms (Roman numerals), CJK Compatibility and Enclosed CJK Letters and Months (certain abbreviations), Latin (certain digraphs), Currency Symbols, Mathematical Alphanumeric Symbols, and Alphabetic Presentation Forms (typographic ligatures). Accented characters Two names which differ only in an accent on one character may look very similar, particularly when the substitution involves the dotted letter i; the tittle (dot) on the i can be replaced with a diacritic (such as a grave accent or acute accent; both ì and í are included in most standard character sets and fonts) that can only be detected with close inspection. In most top-level domain registries, wíkipedia.tld (xn--wkipedia-c2a.tld) and wikipedia.tld are two different names which may be held by different registrants. One exception is .ca, where reserving the plain-ASCII version of the domain prevents another registrant from claiming an accented version of the same name. Non-displayable characters Unicode includes many characters which are not displayed by default, such as the zero-width space. In general, ICANN prohibits any domain with these characters from being registered, regardless of TLD. Known homograph attacks In 2011, an unknown source (registering under the name "Completely Anonymous") registered a domain name homographic to television station KBOI-TV's to create a fake news website. The sole purpose of the site was to spread an April Fool's Day joke regarding the Governor of Idaho issuing a supposed ban on the sale of music by Justin Bieber. In September 2017, security researcher Ankit Anubhav discovered an IDN homograph attack where the attackers registered adoḅe.com to deliver the Betabot trojan. Defending against the attack Client-side mitigation The simplest defense is for web browsers not to support IDNA or other similar mechanisms, or for users to turn off whatever support their browsers have. That could mean blocking access to IDNA sites, but generally browsers permit access and just display IDNs in Punycode. Either way, this amounts to abandoning non-ASCII domain names. Google Chrome versions 51 and later use an algorithm similar to the one used by Firefox. Previous versions display an IDN only if all of its characters belong to one (and only one) of the user's preferred languages. Chromium and Chromium-based browsers such as Microsoft Edge (since 2019) and Opera also use the same algorithm. Safari's approach is to render problematic character sets as Punycode. This can be changed by altering the settings in Mac OS X's system files. Mozilla Firefox versions 22 and later display IDNs if either the TLD prevents homograph attacks by restricting which characters can be used in domain names or labels do not mix scripts for different languages. Otherwise IDNs are displayed in Punycode. Internet Explorer versions 7 and later allow IDNs except for labels that mix scripts for different languages. Labels that mix scripts are displayed in Punycode. There are exceptions to locales where ASCII characters are commonly mixed with localized scripts. Internet Explorer 7 was capable of using IDNs, but it imposes restrictions on displaying non-ASCII domain names based on a user-defined list of allowed languages and provides an anti-phishing filter that checks suspicious Web sites against a remote database of known phishing sites. Old Microsoft Edge converts all Unicode into Punycode. As an additional defense, Internet Explorer 7, Firefox 2.0 and above, and Opera 9.10 include phishing filters that attempt to alert users when they visit malicious websites. As of April 2017, several browsers (including Chrome, Firefox and Opera) were displaying IDNs consisting purely of Cyrillic characters normally (not as punycode), allowing spoofing attacks. Chrome tightened IDN restrictions in version 59 to prevent this attack. Browser extensions like No Homo-Graphs are available for Google Chrome and Firefox that check whether the user is visiting a website which is a homograph of another domain from a user-defined list. These methods of defense only extend to within a browser. Homographic URLs that house malicious software can still be distributed, without being displayed as Punycode, through e-mail, social networking or other Web sites without being detected until the user actually clicks the link. While the fake link will show in Punycode when it is clicked, by this point the page has already begun loading into the browser and the malicious software may have already been downloaded onto the computer. Server-side/registry operator mitigation The IDN homographs database is a Python library that allows developers to defend against this using machine learning-based character recognition. ICANN has implemented a policy prohibiting any potential internationalized TLD from choosing letters that could resemble an existing Latin TLD and thus be used for homograph attacks. Proposed IDN TLDs .бг (Bulgaria), .укр (Ukraine) and .ελ (Greece) have been rejected or stalled because of their perceived resemblance to Latin letters. All three (and Serbian .срб and Mongolian .мон) have later been accepted. Three-letter TLD are considered safer than two-letter TLD, since they are harder to match to normal Latin ISO-3166 country domains; although the potential to match new generic domains remains, such generic domains are far more expensive than registering a second- or third-level domain address, making it cost-prohibitive to try to register a homoglyphic TLD for the sole purpose of making fraudulent domains (which itself would draw ICANN scrutiny). The Russian registry operator Coordination Center for TLD RU only accepts Cyrillic names for the top-level domain .рф, forbidding a mix with Latin or Greek characters. However the problem in .com and other gTLDs remains open. See also Internationalized domain name Homoglyph Duplicate characters in Unicode Unicode equivalence Typosquatting References External links Homograph attack generator Phishing with Unicode Domains Internationalized domain names Nonstandard spelling Unicode Deception Obfuscation Web security exploits Orthography
5110353
https://en.wikipedia.org/wiki/Chartwell%20Technology
Chartwell Technology
Chartwell Technology was a gaming software developer focused largely on the online casino market. The company was based in Calgary. History Chartwell Technology was originally founded in British Columbia on 16 December 1987 as Napier Explorations Inc., a mineral exploration company with interests in mineral exploration in British Columbia. In 1989, the name changed to Brockton Resources Inc. In 1992, the company stopped its mineral activities and moved to oil and gas, changing its name to Chartwell Ventures Ltd. In 1998, after acquiring Gateway Technology, a software developer, the company completely changed its business module by shifting to software developing and licensing. Few months later, on 8 December 1998, the name was eventually changed to Chartwell Technology Inc. In 2005, Chartwell acquired MicroPower, an online poker software developer. The company developed a wide range of online casino games including slot games, table games and fixed odds games, mini games, and networked multiplayer baccarat and blackjack. Chartwell Technology had over 130 employees in three countries, Canada, the United Kingdom, and Malta. Chartwell provided gaming software to notable companies and websites to power their online games, such as Betfair, Bwin, Casinos Austria, Bet365, and 888casino. It also partnered with many gaming software developers, such as Macau-based Entertasia and UK-based HoGaming. Previously traded on the Alberta Stock Exchange and the TSX Venture Exchange, in 2004, Chartwell Technology was listed on the Toronto Stock Exchange under the trading symbol TSX:CWH, and delisted in May 2011 after Amaya's acquisition. Acquisition In July 2011, Chartwell Technology was sold to The Stars Group (then Amaya Gaming Group) in a $23 million deal. As a result, it became a wholly owned subsidiary of Amaya Gaming Group Inc., and its name was changed to Amaya Alberta Inc. In 2015, Amaya sold the company—along with other subsidiaries, to Las Vegas-based NYX Gaming Group, which in turn was acquired a few years later by Scientific Games. External links Official website (archived) References Online gambling companies of Canada The Stars Group 1987 establishments in British Columbia
49921579
https://en.wikipedia.org/wiki/Krzysztof%20Cios
Krzysztof Cios
Krzysztof J. Cios (born 1950) is a Professor and Chair of the Department of Computer Science, School of Engineering, Virginia Commonwealth University (VCU), located in Richmond, Virginia. His research is focused on machine learning, data mining, and biomedical informatics. Biography Krzysztof J. Cios, a Polish-American computer scientist, was born in Dębica, Poland. He earned his M.S. degree in electrical engineering and his Ph.D. degree in computer science, both from the AGH University of Science and Technology, Kraków, Poland. Cios also earned his D.Sc. (habilitation) degree from the Polish Academy of Sciences, Warsaw and an M.B.A. degree from the University of Toledo, Ohio. Cios first worked at the Tadeusz Kościuszko University of Technology, Kraków, and also the International Atomic Energy Agency, Vienna, Austria. After moving to the U.S. in the 1980s he worked at the Electrical Engineering and Computer Science Department at the University of Toledo (UT). At UT he co-founded Department of Bioengineering and served as its Acting Chair before moving to Colorado. He also worked as a Visiting Scientist at NASA Glenn Research Center, Cleveland, Ohio. He then served as Chair of the Computer Science and Engineering Department at the University of Colorado Denver where he was instrumental in starting its Ph.D. program in Computer Science and Information Systems and getting its first-ever ABET accreditation. While in Colorado, he consulted for Lockheed Martin and Raytheon companies and was Affiliated Professor of Computer Science at the University of Colorado Boulder, where he was advisor of two doctoral students. He has also worked as a Visiting Professor at the Complutense University of Madrid, Spain, and at the Muroran Institute of Technology, Japan. Currently he serves as Chair of the Computer Science Department at VCU. He started a dual Ph.D. degree program in computer science between VCU and the University of Cordoba (UCO), Spain; students who meet requirements of both universities and perform research at the other university earn two Ph.D. degrees: from VCU and UCO. He also serves as Director of Enterprise Informatics, Office of the Vice President for Research and Innovation. Cios has co-authored three books and over 200 journal and conference papers. His main contributions are in the areas of machine learning, computational neuroscience, and data mining; his book on the latter subject was the first one published in the U.S. (1998) and his article on uniqueness of medical data mining is frequently cited. In the area of neuroinformatics, he co-defined with Kevin Staley (Harvard Medical School) and others a dynamic Synaptic Activity Plasticity Rule, which in contrast to other plasticity rules is continuous and dynamic as it uses actual post-synaptic potential function to modify connection strength between neurons. Some of the machine learning algorithms developed with his students, such as CAIM and ur-CAIM discretization algorithms and rule algorithms for single-instance (DataSqueezer), multiple-instance (mi-DS) and one-class (OneClass-DS) learning, became popular and were implemented in open-source software platforms. During his academic career Cios has served as primary advisor to 15 Ph.D. students and 20 M.S. students. The former work at universities in the U.S. (one as endowed chair of data science), Australia, Thailand and Vietnam, as post-doctoral researchers at NIH, and at companies such as Google, NASA, General Motors, and Procter & Gamble. His research has been funded by the National Institutes of Health, NASA, National Science Foundation, Colorado Institute of Technology, American College of Cardiology, Ohio Aerospace Institute, U.S. Air Force, and NATO. Cios has been the recipient of the Norbert Wiener Outstanding Paper Award (Kybernetes), the University of Toledo Outstanding Faculty Research Award, the Neurocomputing Best Paper Award and the Fulbright Senior Scholar Award. Cios is listed by the Kosciuszko Foundation as one of the eminent scientists of Polish origin and ancestry, is a Foreign Member of the Polish Academy of Arts and Sciences, and a Fellow of the American Institute for Medical and Biological Engineering. Works References External links His academic home page Virginia Commonwealth University faculty American computer scientists 21st-century American engineers Polish emigrants to the United States AGH University of Science and Technology alumni University of Toledo alumni University of Colorado Boulder faculty Complutense University of Madrid faculty Living people Fellows of the American Institute for Medical and Biological Engineering 1950 births
70158852
https://en.wikipedia.org/wiki/MacBook%20Pro%20%28Intel-based%29
MacBook Pro (Intel-based)
The Intel-based MacBook Pros were a line of Macintosh notebook computers introduced in January 2006 by Apple Inc. and discontinued in November 2020. It was the higher-end model of the MacBook family, sitting above the consumer-focused MacBook Air, and was sold with 13-inch to 17-inch screens. The first-generation MacBook Pro used the design of the PowerBook G4, but replaced the PowerPC G4 chips with Intel Core processors, added a webcam, and introduced the MagSafe power connector. The 15-inch model was introduced in January 2006; the 17-inch model in April. Later revisions added Intel Core 2 Duo processors and LED-backlit displays. The second-generation model debuted in October 2008 in 13- and 15-inch variants, with a 17-inch variant added in January 2009. Called the "unibody" model because its case was machined from a single piece of aluminum, it had a thinner flush display, a redesigned trackpad whose entire surface consisted of a single clickable button, and a redesigned keyboard. Updates brought Intel Core i5 and i7 processors and introduced Intel's Thunderbolt. The third-generation MacBook Pro was released in 2012: the 15-inch in June 2012, a 13-inch model in October. It is thinner than its predecessor, made solid-state storage (SSD) standard, added HDMI, and included a high-resolution Retina display. It eliminated Ethernet and FireWire ports and the optical drive. The fourth-generation MacBook Pro, released in October 2016, adopted USB-C for all data ports and power and included a shallower "butterfly"-mechanism keyboard. On all but the base model, the function keys were replaced with a touchscreen strip called the Touch Bar with a Touch ID sensor integrated into the power button. A November 2019 revision to the fourth-generation MacBook Pro introduced the Magic Keyboard, which uses a scissor-switch mechanism. The initial 16-inch model with a screen set in narrower bezels was followed by a 13-inch model in May 2020. First generation (Aluminum), 2006–2008 The original 15-inch MacBook Pro was announced on January 10, 2006, by Steve Jobs at the Macworld Conference & Expo. The 17-inch model was unveiled on April 24, 2006. The first design was largely a carryover from the PowerBook G4, but uses Intel Core CPUs instead of PowerPC G4 chips. The 15-inch MacBook Pro weighs the same as the 15-inch aluminum PowerBook G4, but is deeper, wider, and thinner. Other changes from the PowerBook include a built-in iSight webcam and the inclusion of MagSafe, a magnetic power connector designed to detach easily when yanked. These features were later brought over to the MacBook. The optical drive was shrunk to fit into the slimmer MacBook Pro; it runs slower than the optical drive in the PowerBook G4 and cannot write to dual-layer DVDs. Both the original 15- and 17-inch model MacBook Pro computers come with ExpressCard/34 slots, which replace the PC Card slots found in the PowerBook G4. Initial first-generation 15-inch models retains the two USB 2.0 ports and a FireWire 400 port but drops the FireWire 800, until it was readded in a later revision, the 17-inch models have an additional USB 2.0 port, as well as the FireWire 800 port missing from the initial 15-inch models. All models now included 802.11a/b/g. Later models include support for the draft 2.0 specification of 802.11n and Bluetooth 2.1. Updates Apple refreshed the entire MacBook Pro line on October 24, 2006, to include Intel Core 2 Duo processors. Memory capacity was doubled for each model, up to 2GB for the high-end 15- and 17-inch models. FireWire 800 was added to the 15-inch models and hard drive capacity was also increased. The MacBook Pro line received a second update on June 5, 2007, with new Nvidia Geforce 8600M GT video cards and faster processor options. LED backlighting was added to the 15-inch model's screen, and its weight was reduced from to . Furthermore, the speed of the front-side bus was increased from 667 to 800MHz. On November 1, 2007, Apple added the option of a 2.6GHz Santa Rosa platform Core 2 Duo CPU as well as reconfigured hard drive options. On February 26, 2008, the MacBook Pro line was updated once more. LED backlighting was added as an option for the 17-inch model. Processors were updated to "Penryn" cores, which are built on the 45nm process (65nm "Merom" cores were previously used), and hard drive and memory capacities were increased. Multi-touch capabilities, first introduced with the MacBook Air earlier that year, were brought over to the trackpad. The original case design was discontinued on October 14, 2008, for the 15-inch, and January 6, 2009, for the 17-inch. Reception Some reviewers applauded the MacBook Pro for its doubling or tripling the speed of the PowerBook G4 in some areas. For example, the 3D rendering program Cinema 4D XL was 3.3 times as fast (2.3 times faster), and its boot-up time was more than twice as quick. The MacBook Pro generally outperformed the PowerBook G4 in performance analyzer utility tests XBench and Cinebench. Reviewers lauded the screen's maximum brightness, 67% brighter than the PowerBook G4; the excellent horizontal viewing angles; the matte options; and the bright, crisp, and true colors. Although the screen offered fewer vertical pixels ( in the MacBook Pro instead of in the PowerBook), one reviewer called the screen "nothing less than stellar". Reviewers praised the new MagSafe power adapter, although one reviewer said it disconnected too easily in some instances. They also praised the backlit keyboard, large trackpad, and virtually silent operation of the machine. The new notebook also offered better wireless performance. One reviewer criticized the decision to underclock the ATI Mobility Radeon X1600 graphics card by about 30% its original speed. The notebook was also noted for running hot. Users complained that upgrading system memory was harder than in older Apple notebooks. Since the dimensions for the 15-inch MacBook Pro were tweaked slightly from the 15-inch PowerBook G4, older accessories such as notebook sleeves did not work with the new models. Some users noted a slight flickering when the screen was on lower brightness settings. Apple increased the battery capacity by 10Wh, going from 50 in the PowerBook G4 to 60, but the more powerful Core Duo CPU required more power. Battery life therefore remained about the same as in previous models, at three-plus hours. 2007–2008 model GPU problems Models of the MacBook Pro built from 2007 to early 2008 (15") / late 2008 (17") using the Nvidia 8600M GT chip reportedly exhibited failures in which the GPU die would detach from the chip carrier, or the chip would detach from the logic board. The issue has been mitigated by some users by keeping the notebook cooler by means of less intensive use or alternative fan settings. Apple initially ignored reports, before admitting to the fault and replacing logic boards free of charge for up to 4 years after the purchase date. NVIDIA also confirmed the issue, and previously manufactured replacement GPUs, which some users have replaced themselves. Technical specifications Second generation (Unibody), 2008–2012 On October 14, 2008, in a press event at company headquarters, Apple officials announced a new 15-inch MacBook Pro featuring a "precision aluminum unibody enclosure" and tapered sides similar to those of the MacBook Air. Designers shifted the MacBook Pro's ports to the left side of the case, and moved the optical disc drive slot from the front to the right side, similar to the MacBook. The new MacBook Pro computers had two video cards that the user could switch between: the Nvidia GeForce 9600M GT with either 256 or 512MB of dedicated memory and a GeForce 9400M with 256MB of shared system memory. Although the FireWire 400 port was removed, the FireWire 800 port remained. The DVI port was replaced with a Mini DisplayPort receptacle. The original unibody MacBook Pro came with a user-removable battery; Apple claimed five hours of use, with one reviewer reporting results closer to four hours on a continuous video battery stress test. Apple said that the battery would hold 80% of its charge after 300 recharges. Design The unibody-construction MacBook Pro largely follows the styling of the original aluminum iMac and the MacBook Air and is slightly thinner than its predecessor, albeit wider and deeper due to the widescreen display. The screen is high-gloss, covered by an edge-to-edge reflective glass finish, while an anti-glare matte option is available in the 15- and 17-inch models in which the glass panel is removed. The entire trackpad is usable and acts as a clickable button. The trackpad is also larger than that of the first generation, giving more room for scrolling and multi-touch gestures. When the line was updated in April 2010, inertial scrolling was added, making the scrolling experience much like that of the iPhone and iPad. The keys, which are still backlit, are now identical to those of Apple's now-standard sunken keyboard with separated black keys. The physical screen release latch from the previous generation is replaced with a magnetic one. Updates During the MacWorld Expo keynote on January 6, 2009, Phil Schiller announced a 17-inch MacBook Pro with unibody construction. This version diverged from its 15-inch sibling with an anti-glare "matte" screen option (with the glossy finish standard) and a non user-removable lithium polymer battery. Instead of traditional round cells inside the casing, the lithium-ion polymer batteries are shaped and fitted into each notebook to maximally utilize space. Adaptive charging, which uses a chip to optimize the charge flow to reduce wear and tear, extends the battery's overall life. Battery life for the 17-inch version is quoted at eight hours, with 80 percent of this charge remaining after 1,000 charge-discharge cycles. At Apple's Worldwide Developers Conference (WWDC) on June 8, 2009, it was announced that the 13-inch unibody MacBook would be upgraded and re-branded as a MacBook Pro, leaving only the white polycarbonate MacBook in the MacBook line. It was also announced that the entire MacBook Pro line would use the non-user-removable battery first introduced in the 17-inch MacBook Pro. The updated MacBook Pro 13- and the 15-inch would each have up to a claimed 7 hours of battery life, while the 17-inch would keep its 8-hour capacity. Some sources even reported up to eight hours of battery life for the 13- and 15-inch MacBook Pro computers during casual use, while others reported around six hours. Like the 17-inch MacBook Pro, Apple claims that they will last around 1,000 charging cycles while still containing 80% of their capacity. Graphics card options stayed the same from the previous release, although the 13-inch and the base model 15-inch, came with only the GeForce 9400M GPU. The screens were also improved, gaining a claimed 60 percent greater color gamut. All of these mid-2009 models also included a FireWire 800 port and all except the 17-inch models would receive an SD card slot. The 17-inch model would retain its ExpressCard/34 slot. For the 13-inch MacBook Pro, the Kensington lock slot was moved to the right side of the chassis. In August 2009, Apple extended the "matte" anti-glare display option to the 15-inch MacBook Pro. On April 13, 2010, Intel Core i5 and Core i7 CPUs were introduced in the 15- and 17-inch models, while the 13-inch retained the Core 2 Duo with a speed increase. The power brick was redesigned and a high-resolution display (of ) was announced as an option for the 15-inch models. The 13-inch gained an integrated Nvidia GeForce 320M graphics processing unit (GPU) with 256MB of shared memory, while the 15- and 17-inch models were upgraded to the GeForce GT 330M, with either 256 or 512MB of dedicated memory. The 15- and 17-inch models also have an integrated Intel GPU that is built into the Core i5 and i7 processors. The 15-inch model also gained . Save for a third USB 2.0 slot, all the ports on the 17-inch MacBook Pro are the same in type and number as on the 15-inch version. All models come with 4GB of system memory that is upgradeable to 8GB. Battery life was also extended further in this update, to an estimated 10 hours for the 13-inch and 8–9 hours on the 15- and 17-inch MacBook Pro computers. This was achieved through both greater power efficiency and adding more battery capacity. One reviewer reported about 6 hours of battery life through a continuous video battery stress test in the 15-inch and another, who called the battery life "unbeatable", reported nearer to 8 in the 13-inch through their "highly demanding battery drain test". Thunderbolt technology, Sandy Bridge dual-core Intel Core i5 and i7 (on the 13-inch model) or quad-core i7 (on the 15- and 17-inch models) processors, and a high definition FaceTime camera were added on February 24, 2011. Intel HD Graphics 3000 come integrated with the CPU, while the 15- and 17-inch models also utilize AMD Radeon HD 6490M and Radeon HD 6750M graphics cards. Later editions of these models, following the release of OS X Lion, replaced the Expose (F3) key with a Mission Control key, and the Dashboard (F4) key with a Launchpad key. The chassis bottoms are also engraved differently from the 2010 models. The Thunderbolt serial bus platform can achieve speeds of up to 10Gbit/s, which is up to twice as fast as the USB 3.0 specification, 20 times faster than the USB 2.0 specification, and up to 12 times faster than FireWire 800. Apple says that Thunderbolt can be used to drive displays or to transfer large quantities of data in a short amount of time. On June 11, 2012, Apple showcased its upgraded Mac notebooks, OS X Mountain Lion, and iOS 6 at the Worldwide Developers Conference (WWDC) in San Francisco. The new MacBook Pro models were updated with Ivy Bridge processors and USB 3.0 ports, and the default RAM on premium models was increased to 8GB. Following this announcement, the 17-inch model was discontinued. After a media event on October 22, 2013, Apple discontinued all second-generation MacBook Pro computers except for the entry-level 2.5GHz 13-inch model. Apple discontinued the 13-inch second-generation MacBook Pro on October 27, 2016. Prior to its discontinuation it was Apple's only product to still include an optical drive and a FireWire port, and only notebook with a hard disk drive and Ethernet port. It is also the only MacBook Pro to support 9 versions of MacOS X/MacOS, from Mac OS X Lion 10.7 through MacOS Catalina 10.15. Reception Some reviewers praised the new notebook's performance and compact size, the quality of the screen, and sturdy unibody build, which allowed easier upgrading of internal components as compared to the original models. Some reviewers also noted that the new MacBook Pro ran more quietly and at cooler temperatures than first-generation machines. Others, however, criticized the amount of heat generated by the new design. The Reviewers lamented the loss of a matte screen option for the 2008 unibody MacBook Pro, noting the reflectiveness of the screen in sunlight, even when its brightness was turned all the way up. CNET's Dan Ackerman commented of the mid-2009 models: "According to Apple, the new display offers a wider color gamut, and the screen certainly looks bright and colorful, but we wish the same matte-screen option offered on the 17-inch MacBook Pro was available across the line... While the LED screen means a thinner lid and some battery life benefits, the edge-to-edge glass covering the entire display panel grabs stray light rays with ease, making the glossy screen hard to see in some lighting conditions." By 2011, matte screens were offered for both the 15" and 17" models. Furthermore, the addition of Mini DisplayPort instead of the more popular HDMI was criticized. The relatively low number of ports and lower end technical specifications when compared to similarly priced laptops from other brands were also bemoaned. Laptop Magazines Michael Prospero praised the 2010 15-inch model's display, calling it "bright and crisp". He further commented, "While reflections from the glossy display weren't overwhelming, it's also nice to know there's an antiglare option—though only for the higher resolution display. Still, colors were bright, blacks were deep and dark, and viewing angles were excellent both vertically and horizontally." He also lauded the quality of the iSight webcam, the responsiveness of the trackpad, the microphone and speakers, as well as the performance of the new CPUs for the 15" model and the long battery life. Complaints included the price of the notebook, the low number of USB ports, and the lack of HDMI. CNET praised the automatic graphics switching features of the 15- and 17-inch 2010 models as well as the graphics cards themselves. Acclaim was also given to the Core i5 and i7 CPUs, the multi-touch trackpad, and the addition of audio capabilities to the Mini DisplayPort video output. They also called for the addition of HDMI and the Blu-ray optical disc format, saying that most other computers in the MacBook Pro's price range possessed these features. CNET also criticized the option of a higher-resolution screen in the 15-inch model, saying that "the higher-resolution screen should be included by default." Technical specifications {| class="wikitable mw-collapsible mw-collapsed" style="width:100%; font-size:90%; text-align:center" |- ! colspan="9" | Comparison of second-generation MacBook Pro models |- ! Component ! colspan="3" | Intel Core 2 Duo ! Intel Core 2 Duo, Core i5, Core i7 ! colspan="4" | Intel Core i5, Core i7 |- ! style="background:#gray; width:10%"| Model ! style="background:#FF9999; width:10%"| Late 2008 ! style="background:#FF9999; width:10%"| Early 2009 ! style="background:#FF9999; width:10%"| Mid 2009 ! style="background:#FF9999; width:10%"| Mid 2010 ! style="background:#FF9999; width:10%" | Early 2011 ! style="background:#FF9999; width:10%" | Late 2011 ! style="background:#ffDDDD; width:10%" | Mid 2012 ! style="background:#FF9999; width:10%" | Mid 2012 |- | Release date | October 14, 2008 | January 6, 2009 (17")March 3, 2009 (15") | June 8, 2009 | April 13, 2010 | February 24, 2011 | October 24, 2011 | colspan="2" | June 11, 2012 |- | Apple order number(s) | MB470*/A or MB471*/A | MB470*/A, MC026*/A, MB604*/A | MB990*/A, MB991*/A, MC118*/A, MB985*/A, MB986*/A, MC226*/A | MC374*/A, MC375*/A, MC371*/A, MC372*/A, MC373*/A, MC024*/A | MC700*/A, MC724*/A, MC721*/A, MC723*/A, MC725*/A | MD311*/A, MD313*/A, MD314*/A, MD318*/A, MD322*/A, MD385*/A | MD101*/A, MD102*/A | MD103*/A, MD104*/A |- | Model number(s) | A1286 (15") | A1286 (15"), A1297 (17") | colspan="4" | A1278 (13"), A1286 (15"), A1297 (17") | A1278 (13") | A1286 (15") |- | Model identifier(s) | MacBookPro5,1 (15") | MacBookPro5,1(15")MacBookPro5,2(17") | MacBookPro5,1(15")MacBookPro5,2(17")MacBookPro5,3(15")MacBookPro5,4(15")MacBookPro5,5(13") | MacBookPro6,1(17")MacBookPro6,2(15")MacBookPro7,1(13") | colspan="2" | MacBookPro8,1 (13")MacBookPro8,2 (15")MacBookPro8,3 (17") | MacBookPro9,2 (13") | MacBookPro9,1 (15") |- | rowspan="3" | LED-backlit widescreen glossy display (16:10) | colspan="2" | colspan="5" | 13.3", | |- | colspan="2" | 15.4", | 15.4", Optional matte screen | colspan="3" | 15.4", Optional (glossy or matte) | | 15.4", Optional (glossy or matte) |- | | colspan="5" | 17", Optional matte screen | colspan="2" |- | Video camera | colspan="4" | iSight (480p) | colspan="4" | FaceTime HD (720p) |- | rowspan=2| Processor | colspan=2 | 2.26GHz (P8400) or 2.53GHz (P8700) Intel Core 2 Duo Penryn with 3MB on-chip L2 cache | 2.4GHz (P8600) or 2.66GHz (P8800) Intel Core 2 Duo Penryn with 3MB on-chip L2 cache | 2.3GHz (2415M) Intel Core i5 Sandy Bridge with 3MB on-chip L3 cache or 2.7GHz (2620M) Intel Core i7 Sandy Bridge with 4MB on-chip L3 cache | 2.4GHz (2435M) Intel Core i5 Sandy Bridge with 3MB on-chip L3 cache or 2.8GHz (2640M) Intel Core i7 Sandy Bridge with 4MB on-chip L3 cache | 2.5GHz (3210M) Intel Core i5 Ivy Bridge with 3MB on-chip L3 cache or 2.9GHz (3520M) Intel Core i7 Ivy Bridge with 4MB on-chip L3 cache | |- | 2.4GHz (P8600) with 3MB on-chip L2 cache or 2.53GHz (T9400) Intel Core 2 Duo Penryn with 6MB on-chip L2 cacheOptional 2.8 GHz (T9600) with 6MB on-chip L2 cache | 2.4GHz (P8600) with 3MB on-chip L2 cache (15" only) or 2.53GHz (T9400) with 6MB on-chip L2 cache (15" only) or 2.66GHz (T9550) (17" only) Intel Core 2 Duo Penryn with 6MB on-chip L2 cache Optional 2.8 GHz (T9600) with 6MB on-chip L2 cache (15" only) or 2.93GHz (T9800) with 6MB on-chip L2 cache (17" only)'| 2.53GHz (P8700) or 2.66GHz (P8800) (15" only) with 3MB on-chip L2 cache or 2.8GHz (T9600) (15" and 17" only) Intel Core 2 Duo Penryn with 6MB on-chip L2 cacheOptional 3.06 GHz (T9900) with 6MB on-chip L2 cache (15" and 17" only) | 2.4GHz (520M) (15" only) or 2.53GHz (540M) (15" and 17" only) with 3MB on-chip L3 cache or 2.66GHz (620M) (15" and 17" only) Intel Core i7 Arrandale with 4MB on-chip L3 cacheOptional 2.8 GHz (640M) with 4MB on-chip L3 cache (15" and 17" only) | 2.0GHz quad-core (2635QM) (15" only) or 2.2GHz quad-core (2720QM) (15" and 17" only) Intel Core i7 Sandy Bridge with 6MB on-chip L3 cacheOptional 2.3 GHz (2820QM) with 8MB on-chip L3 cache (15" and 17" only) | 2.2GHz quad-core (2675QM) (15" only) or 2.4GHz quad-core (2760QM) (15" and 17" only) Intel Core i7 Sandy Bridge with 6MB on-chip L3 cacheOptional 2.5 GHz (2860QM) with 8MB on-chip L3 cache (15" and 17" only) | | 2.3GHz quad-core (3615QM) (15" only) or 2.6GHz quad-core (3720QM) (15" only) Intel Core i7 Ivy Bridge with 6MB on-chip L3 cacheOptional 2.7 GHz (3820QM) with 8MB on-chip L3 cache (15" only) |- | rowspan="2" | System bus| colspan="2" | colspan="2" | 1,066MHz front-side bus (13") | rowspan="2" colspan="4" | DMI 5 GT/s |- | 1,066MHz front-side bus | colspan="2" | 1,066MHz front-side bus (15" and 17") | Intel DMI2.5 GT/s (15" and 17") |- | rowspan=2| Memory(two slots)| 2GB (two 1GB) or 4GB (two 2GB)Expandable to 4GB by default, expandable to 8GB with the latest EFI update | 4GB (two 2GB)Expandable to 8GB. 2.66 and 2.93GHz models expandable to 8GB | 2GB (two 1GB) or 4GB (two 2GB)Expandable to 8GB | 4GB (two 2GB)Expandable to 8GB16GB on 13" models | colspan="2" | 4GB (two 2GB)Expandable to 16GB | colspan="2" | 4GB (two 2GB) or 8GB (two 4GB)Expandable to 16GB |- | colspan="4" | 1066MHz PC3-8500 DDR3 SDRAM | colspan="2" | 1333MHz PC3-10600 1.5V DDR3 SDRAMExpandable to 16GB of 1600MHz PC3-12800 DDR3 SDRAM | colspan="2" | 1600MHz PC3-12800 1.35 V DDR3 SDRAM |- | rowspan="2" | Graphics| colspan="2" | Nvidia GeForce 9400M with 256MB of DDR3 SDRAM shared with main memory (13" and some 15" models) | Nvidia GeForce 320M with 256MB of DDR3 SDRAM shared with main memory (13" models only) | colspan="2" | Intel HD Graphics 3000 with 384MB (512MB with 8GB of RAM installed) DDR3 SDRAM shared with main memory (13" models only) | Intel HD Graphics 4000 with DDR3 SDRAM shared with main memory | |- | colspan="2" | Nvidia GeForce 9400M with 256MB of DDR3 SDRAM shared with main memory and Nvidia GeForce 9600M GT with 256 or 512MB of GDDR3 SDRAMCan switch between the two (but cannot use both) | Nvidia GeForce 9400M with 256MB of DDR3 SDRAM shared with main memory and Nvidia GeForce 9600M GT with 256MB or 512MB of GDDR3 SDRAM (some 15" or 17" models)Can switch between the two (but cannot use both) | Intel HD Graphics with 256MB of DDR3 SDRAM shared with main memory and Nvidia GeForce GT 330M with 256MB or 512MB of GDDR3 SDRAM (15" and 17" models)Automatically switches between graphics hardware when running OS X | Intel HD Graphics 3000 with 384MB DDR3 SDRAM shared with main memory(15" and 17" models) and AMD Radeon HD 6490M with 256MB GDDR5 memory (15" models) or AMD Radeon HD 6750M with 1GB GDDR5 memory (15" and 17" models)Automatically switches between graphics hardware when running OS X | Intel HD Graphics 3000 with 384MB DDR3 SDRAM shared with main memory(15" and 17" models) and AMD Radeon HD 6750M with 512MB GDDR5 memory (15" models) or AMD Radeon HD 6770M with 1GB GDDR5 memory (15" and 17" models)Automatically switches between graphics hardware when running OS X | | Intel HD Graphics 4000 with DDR3 SDRAM shared with main memory and Nvidia GeForce GT 650M with 512MB GDDR5 memory (base 15" model) or 1GB GDDR5 memoryAutomatically switches between graphics hardware when running OS X |- | rowspan="2" | Storage| 250 or 320GB SATA at 5,400RPMOptional 250 or 320GB at 7,200 RPM, 128GB SSD | 250 or 320GB SATA at 5,400RPMOptional 250 or 320GB at 7,200 RPM, 128 or 256GB SSD | 160, 250, 320 or 500GB SATA at 5,400RPMOptional 320 or 500GB at 5,400 or 7,200 RPM (15" and 17" only) or 128 or 256GB SSD | 250, 320, or 500GB SATA at 5,400RPM.Optional 320 or 500GB at 5,400 RPM or 7,200 RPM, or 128, 256, or 512GB SSD. | 320 (13" only), 500, or 750GB SATA at 5,400RPM.Optional 500 or 750GB at 5,400 RPM or 500GB at 7,200 RPM (15" and 17" only), or 128, 256, or 512GB SSD | 500 or 750GB SATA at 5,400RPM (13" and base 15" models) or 750GB SATA at 5,400RPM (high-end 15" and 17" models)Optional 750GB at 5,400 or 7,200 RPM (15" and 17" only), or 128, 256, or 512GB SSD | colspan="2" | 500 or 750GB SATA at 5,400RPM Optional 750GB at 5,400 or 7,200 RPM or 1TB at 5,400 RPM or 128, 256, 512GB SSD |- | colspan="4" | SATA 3 Gbit/s | colspan="4" | SATA 6 Gbit/s |- | Optical disc drive| colspan="8" | SuperDrive: 4× DVD±R DL writes, 8× DVD+/−R read/write, 8× DVD+RW writes, 6× DVD-RW writes, 24× CD-R, and 16× CD-RW recording |- | rowspan="3" | Connectivity| colspan="4" | Integrated AirPort Extreme (802.11a/b/g/draft-n) (Broadcom BCM4322 chipset, up to 300Mbit/s) | colspan="4" | Integrated AirPort Extreme (802.11a/b/g/n) (Broadcom BCM4331 chipset, up to 450Mbit/s) |- | colspan="6" | Bluetooth 2.1 + EDR | colspan="2" | Bluetooth 4.0 |- | colspan="8" | Gigabit Ethernet |- | rowspan="4" | Peripheral connections| colspan="2" | ExpressCard/34 | colspan="4" | SDXC card slot (13" and 15") or ExpressCard/34 (17") | colspan="2" | SDXC card slot |- | colspan="6" | USB 2.0 (two ports on 13" and 15", three ports on 17") | colspan="2" | USB 3.0 (two ports) |- | colspan="3" | Mini DisplayPort (without audio support) | Mini DisplayPort (with audio support) | colspan="4" | Thunderbolt port |- | colspan="8" | Firewire 800 |- | rowspan="2" | Audio| colspan="8" | Built-in stereo speakers |- | colspan="8" | Audio line-in/out |- | Minimum operating system| colspan="3" | Mac OS X 10.5 Leopard | colspan="2" | Mac OS X 10.6 Snow Leopard | colspan="3" | Mac OS X 10.7 Lion |- | Latest release operating system| colspan="3" | OS X 10.11 El Capitan | colspan="3" | macOS 10.13 High Sierra | colspan="2" | macOS 10.15 Catalina |- | rowspan="3" | Battery (lithium polymer, non-removable except in original 15") | colspan="2" | 58 Wh (13") | colspan="4" | 63.5 Wh (13") | |- | colspan="2" | 50 Wh removable lithium-polymer (15") | 73 Wh (15") | colspan="3" | 77.5 Wh (15") | | 77.5 Wh (15") |- | | colspan="5" | 95 Wh (17") | colspan="2" |- | rowspan="3" | Weight| colspan="2" | colspan="5" | (13") | |- | colspan="3" | (15") | colspan="3" | (15") | | (15") |- | | colspan="5" | (17") | colspan="2" |- | rowspan="3" | Dimensions (widthdepththickness)''' | colspan="2" | colspan="5" | (13")| |- | colspan="6" | (15")| | (15")|- | | colspan="5" | (17")| colspan="2" |} Since the RAM and the hard drive on some generations of MacBook Pro are user-serviceable parts, there are aftermarket modifications to enhance the system with up to 16GB of DDR3-1600 RAM (although maximum capacity and frequency depend on the hardware in question), 7200RPM hard drives or third-party SSDs. A third-party caddy was also made, allowing the internal optical drive to be replaced with a second internal SATA 2.5-inch hard drive. Early and late 2011 model GPU problems Early and late 2011 models with a GPU; 15" & 17"; reportedly suffer from manufacturing problems leading to overheating, graphical problems, and eventually complete GPU and logic board failure. A similar but nonidentical problem affected iMac GPUs which were later recalled by Apple. The problem was covered by many articles in Mac-focused magazines, starting late 2013 throughout 2014. In August 2014 the law firm Whitfield Bryson & Mason LLP had begun investigating the problem to determine if any legal claim exists. On October 28, 2014, the firm announced that it has filed a class-action lawsuit in a California federal court against Apple. The lawsuit will cover residents residing in both California and Florida who have purchased a 2011 MacBook Pro notebook with an AMD graphics card. The firm is also investigating similar cases across the United States. On February 20, 2015, Apple instituted the This "will repair affected MacBook Pro systems, free of charge". The program covered affected MacBook Pro models until December 31, 2016, or four years from original date of sale. Third generation (Retina), 2012–2016 On June 11, 2012, at the Apple Worldwide Developers Conference in San Francisco, Apple introduced the third generation MacBook Pro, the 15-inch in June 2012, a 13-inch model in October, marketed as the "MacBook Pro with Retina display" to differentiate it from the previous model. The new model includes Intel's third-generation Core i7 processors (Ivy Bridge microarchitecture). It made solid-state storage (SSD) standard, upgraded to USB 3.0, added an additional Thunderbolt port, added HDMI, and included a high-resolution Retina display. The 15-inch model is 25% thinner than its predecessor. The model name is no longer placed at the bottom of the screen bezel; instead, it is found on the underside of the chassis, similar to an iOS device and is the first Macintosh notebook to not have its model name visible during normal use. It eliminated Ethernet, FireWire 800 ports, but Thunderbolt adapters were available for purchase,, Kensington lock slot, the battery indicator button and light on the side of the chassis, and the optical drive, being the first professional notebook since the PowerBook 2400c, but brought a new MagSafe port, dubbed the "MagSafe 2". Apple also claims improved speakers and microphones and a new system for cooling the notebook with improved fans. The MacBook Pro with Retina display was one of only two Macs that feature a built-in HDMI port (the other being the Mac Mini). Apple introduced a 13-inch version on October 23, 2012, with specifications similar but slightly inferior to the 15-inch version's, such as less powerful processors. The Retina models also have fewer user-accessible upgrade or replacement options than previous MacBooks. Unlike in previous generations, the memory is soldered onto the logic board and is therefore not upgradable. The solid state drive is not soldered and can be replaced by users, although it has a proprietary connector and form factor. The battery is glued into place; attempts to remove it may destroy the battery and/or trackpad. The entire case uses proprietary pentalobe screws and cannot be disassembled with standard tools. While the battery is glued in, recycling companies have stated that the design is only "mildly inconvenient" and does not hamper the recycling process. In any case, Apple offers a free recycling service via any of their stores and has no problem with separating the materials. On February 13, 2013, Apple announced updated prices and processors and increased the memory of the high-end 15-inch model to 16GB. On October 22, 2013, Apple updated the line with Intel's Haswell processors and Iris Graphics, 802.11ac Wi-Fi, Thunderbolt 2, and PCIe-based flash storage. The chassis of the 13-inch version was slightly slimmed to to match the 15-inch model. The lower-end 15-inch model only included integrated graphics while the higher-end model continued to include a discrete Nvidia graphics card in addition to integrated graphics. Support for 4K video output via HDMI was added but limited the maximum number of external displays from three to two. On July 29, 2014, Apple announced new models with updated prices and processors. On March 9, 2015, the 13-inch model was updated with Intel Broadwell processors, Iris 6100 graphics, faster flash storage (based on PCIe 2.0 × 4 technology), faster RAM (upgraded from 1600MHZ to 1866MHZ), increased battery life (extended to 10 hours), and a Force Touch trackpad. On May 19, 2015, 15-inch model added Force Touch and changed the GPU to AMD Radeon R9 M370X, SSD based on PCIe 3.0 × 4 technology, the battery life was extended to 9 hours, and the rest of the configuration remained unchanged. The higher-end 15-inch model also added support for dual-cable output to displays. The 15-inch models were released with the same Intel Haswell processors and Iris Pro graphics as the 2014 models due to a delay in shipment of newer Broadwell quad-core processors. Apple continued to sell the 2015 15-inch model until July 2018. Reception The Retina MacBook Pro received positive reviews of the Retina Display, flash storage and power. It was criticized, however, for its high price and lack of an Ethernet port and optical drive. Roman Loyola of Macworld said that the Retina MacBook Pro was "groundbreaking" and made people "rethink how they use technology". He praised the inclusion of USB 3.0 and the slimmer body. Dan Ackerman of CNET commented "I've previously called the 15-inch MacBook Pro one of the most universally useful all-around laptops you can buy. This new version adds to that with HDMI, faster ports, and more portability. But it also subtracts from that with its exclusion of an optical drive and Ethernet port, plus its very high starting price. The Pro and Retina Pro are clearly two laptops designed for two different users, and with the exception of all-day commuters who need something closer to a MacBook Air or ultrabook, one of the two branches of the MacBook Pro family tree is still probably the most universally useful laptop you can buy." Joel Santo Domingo of PC Magazine gave the MacBook Pro an "Editor's Choice" rating. He praised its "brilliant Retina display", the thin design, port selection and speedy storage, and highlighted the expandability via Thunderbolt ports which support up to seven devices each. David Pogue of The New York Times praised the 15-inch model's screen, keyboard, sound, start-up time, cosmetics, battery life, storage, and RAM capacity. They criticized the lack of a SuperDrive, pricing, and the MagSafe 2 power connector's lack of backwards compatibility with the older MagSafe design. The Retina Display on the MacBook Pro have been criticized for "image retention", specifically for displays manufactured by LG. Many users also complained the anti-reflective coating on their screens could wear off easily, which is an issue known as "staingate". In 2017, one year after the introduction of the fourth generation of the MacBook Pro, the original lead developer of Tumblr Marco Arment wrote an evocative article in which he declared the Retina MacBook Pro the best laptop ever made. The sentiment was shared by many users of various social platforms. Repairability and environmental concerns Apple was criticized for gluing the battery into the case, making it harder to be recycled (ease of disassembly is an EPEAT criterion), but some recycling companies have stated that the design is only "mildly inconvenient" and does not hamper the recycling process. Greenpeace spokesman Casey Harrell said Apple "has pitted design against the environment—and chosen design. They're making a big bet that people don't care, but recycling is a big issue." Wired also criticized Apple's recyclability claims in 2012: "[t]he design may well be 'highly recyclable aluminum and glass'—but my friends in the electronics recycling industry tell me they have no way of recycling aluminum that has glass glued to it like Apple did with both this machine and the recent iPad." Battery problems In June 2019, Apple announced a worldwide recall for certain 2015 15" MacBook Pro computers after receiving at least 26 reports of batteries becoming hot enough to produce smoke and inflict minor burns or property damage. The problem affected some 432,000 computers, mostly sold between September 2015 and February 2017. The company asked customers to stop using their computers until Apple could replace the batteries. In September 2019, India's Directorate General of Civil Aviation said MacBook Pro computers could dangerously overheat, leading the national carrier Air India to ban the model on its flights. Technical specifications Fourth generation (Touch Bar), 2016–2020 Apple unveiled fourth-generation 13- and 15-inch MacBook Pro models during a press event at their headquarters on October 27, 2016. All models, except for the baseline 13-inch model, featured the Touch Bar, a new multi-touch-enabled OLED strip built into the top of the keyboard in place of the function keys. The Touch Bar is abutted on its right by a sapphire-glass button that doubles as a Touch ID sensor and a power button. The models also introduced a "second-generation" butterfly-mechanism keyboard whose keys have more travel than the first iteration in the Retina MacBook. The 13-inch model has a trackpad that is 46% larger than its predecessor while the 15-inch model has a trackpad twice as large as the previous generation. All ports have been replaced with either two or four combination Thunderbolt 3 ports that support USB-C 3.1 Gen 2 and dual DisplayPort 1.2 signals, any of which can be used for charging. The MacBook Pro is incompatible with some older Thunderbolt 3-certified peripherals, including Intel's own reference design for Thunderbolt 3 devices. Furthermore, macOS on MacBook Pro blacklists (prevents from working) certain classes of Thunderbolt 3-compatible devices. Support for Thunderbolt 3 external graphics processing units (eGPU) was added in macOS High Sierra 10.13.4. Devices using HDMI, previous-generation Thunderbolt, and USB need an adapter to connect to the MacBook Pro. The models come with a 3.5mm headphone jack; the TOSLINK functionality of older-generation MacBook Pro computers has been removed. Other updates to the MacBook Pro include dual- and quad-core Intel "Skylake" Core i5 and i7 processors, improved graphics, and displays that offer a 25% wider color gamut, 67% more brightness, and 67% more contrast. All versions can output to a 5K display; the 15-inch models can drive two such displays. The 15-inch models include a discrete Radeon Pro 450, 455 or 460 graphics card in addition to the integrated Intel graphics. The base 13-inch model has function keys instead of the Touch Bar, and just two USB-C ports. The flash storage in the Touch Bar models is soldered to the logic board and is not upgradeable, while in the 13-inch model without Touch Bar, it is removable, but difficult to replace, as it is a proprietary format of SSD storage. On June 5, 2017, Apple updated the line with Intel Kaby Lake processors and newer graphics cards. A 128GB storage option was added for the base 13-inch model, down from the base 256GB storage. New symbols were introduced to the control and option keys. On July 12, 2018, Apple updated the Touch Bar models with Intel Coffee Lake quad-core processors in 13-inch models and six-core processors in 15-inch models, updated graphics cards, third-generation butterfly keyboards that introduced new symbols for the control and option keys, Bluetooth 5, T2 SoC Chip, True Tone display technology, and larger-capacity batteries. The 15-inch model can also be configured with up to 4TB of storage, 32GB of DDR4 memory and a Core i9 processor. In late November the higher-end 15-inch model could be configured with Radeon Pro Vega graphics. On May 21, 2019, Apple announced updated Touch Bar models with newer processors, with an eight-core Core i9 standard for the higher-end 15-inch model, and an updated keyboard manufactured with "new materials" across the line. On July 9, 2019, Apple updated the 13-inch model with two Thunderbolt ports with newer quad-core eighth-generation processors and Intel Iris Plus graphics, True Tone display technology, and replaced the function keys with the Touch Bar. macOS Catalina added support for Dolby Atmos, Dolby Vision, and HDR10 on 2018 and newer models. macOS Catalina 10.15.2 added support for 6016x3384 output on 15-inch 2018 and newer models to run the Pro Display XDR at full resolution. The 2019 MacBook Pro was the final model that could run macOS Mojave 10.14, the final MacOS version that can run 32-bit applications such as Microsoft Office for Mac 2011. Design and usability The fourth-generation MacBook Pro follows the design of the previous two generations, with an all-metal unibody enclosure and separated black keys. A few of the apparent design changes include a thinner chassis and screen bezel, a larger trackpad, the OLED Touch Bar, and the shallower butterfly-mechanism keyboard with less key separation than the previous models. The speaker grilles have been relocated to the sides of the keyboard on the 13-inch variant. Tear downs show that the speaker grilles on the 13-inch model with Touch Bar are "largely cosmetic", and that sound output mostly comes through the side vents. The fourth generation MacBook Pro comes in two finishes, the traditional silver color and a darker "space gray" color. The MacBook Pro model name returns to the bottom of the screen bezel in Apple's San Francisco font after being absent from the second generation with Retina display. As with the Retina MacBook, the new models replace the backlit white Apple logo on the rear of the screen, a feature dating back to the 1999 PowerBook G3, with a glossy metal version. MagSafe, a magnetic charging connector, has been replaced with USB-C charging. Unlike MagSafe, which provided an indicator light within the user's field of view to indicate the device's charging status, the USB-C charger has no visual indicator. Instead, the MacBook Pro emits a chime when connected to power. The Macintosh startup chime that has been used since the first Macintosh in 1984 is now disabled by default. The notebook now boots automatically when the lid is opened. Battery life The battery life of the new models also got a mixed reception, with outlets reporting inconsistent battery life and inaccurate estimates of time remaining on battery by the operating system. After the latter reports, Apple used a macOS update to hide the display of estimated battery time. Consumer Reports did not initially recommend the 2016 MacBook Pro models, citing inconsistent and unpredictable battery life in its lab testing (which involves the consecutive loading of multiple websites). However, Apple and Consumer Reports found that the results had been affected by a bug caused by disabling caching in Safari's developer tools. Consumer Reports performed the tests again with a patched macOS, and retracted its original assessment. Repairability iFixit scored the models 1 out of 10 for repairability, noting that memory, the processor, and flash storage are soldered to the logic board, while the battery is glued to the case. The entire assembly uses proprietary pentalobe screws and cannot be disassembled with standard tools. Keyboard reliability A report by AppleInsider has claimed that the updated "Butterfly" keyboard fails twice as often as previous models, often due to particles stuck beneath the keys. Repairs for stuck keys have been estimated to cost more than $700. In May 2018, two class action lawsuits were filed against Apple regarding the keyboard problem; one alleged a "constant threat of nonresponsive keys and accompanying keyboard failure" and accusing Apple of not alerting consumers to the problem. In June 2018, Apple announced a Service Program to "service eligible MacBook and MacBook Pro keyboards, free of charge". The 2018 models added a membrane underneath keys to prevent malfunction from dust. As of early 2019, there were reports of problems with the same type of keyboards in the 2018 MacBook Air. In May 2019, Apple modified the keyboard for the fourth time and promised that any MacBook keyboard with butterfly switches would be repaired or replaced free of charge for a period of four years after the date of sale. Thermal throttling PC Magazine said "the Core i9 processor Apple chose to use inside the MacBook Pro (i9-8950K) has a base clock frequency of 2.9GHz, which is capable of bursting up to 4.8GHz when necessary. However, testing carried out by YouTuber Dave Lee showed that the Core i9 couldn't even maintain 2.9GHz, let alone 4.8GHz. And it ended up running at 2.2GHz due to the heat generated inside the chassis forcing it to throttle. Lee found the 2018 i9 MacBook Pro was slower than the 2017 MacBook Pro and stated, "This isn't a problem with Intel's Core i9, it's Apple's thermal solution." When Lee put the i9 MacBook Pro inside a freezer, the render times were over 30% faster. On July 24, 2018, Apple released a software fix for the new 2018 MacBook Pro computers which addressed the thermal throttling problem. Apple said "there is a missing digital key in the firmware that impacts the thermal management system and could drive clock speeds down under heavy thermal loads on the new MacBook Pro". Other problems A "limited number" of 13-inch MacBook Pro units without Touch Bar, manufactured between October 2016 and October 2017, saw the built-in battery swell. Apple created a free replacement program for eligible units. A "limited number" of 128 and 256GB solid-state drives used in 13-inch MacBook Pro (non-Touch Bar) units can lose data and fail. 13-inch MacBook Pro units with affected drives were sold between June 2017 and June 2018. This resulted in Apple launching a repair program for those affected – the repair involves the update of firmware. Some users are reporting kernel panics on 2018 models, because of the T2 chip. Apple is already aware of the problem and performing an investigation. There are also user reports about the speaker crackling problems on the 2018 models. Users have reported malfunctioning display cables, causing uneven lighting at the bottom of the screen and ultimately display failure. Customers of Apple have named this issue "Flexgate". The problem has been tracked to a cable, stressed from opening and closing the notebook. The entire display needs to be replaced in affected units. In May 2019 Apple initiated a program to replace the display on affected 13-inch models made in 2016 for free, and the cable on the 2018 models and onwards was made 2mm longer than on prior models, thus reducing the likelihood of display failure. Apple has been criticized for not extending the replacement program to the 15-inch models which are also affected by this issue. Reception The fourth-generation MacBook Pro received mixed reviews. The display, build quality, and audio quality were praised but many complained about the butterfly keyboard; the little-used Touch Bar; and the absence of USB-A ports, HDMI port, and SD card slot.Ars Technica noted that the second-generation keyboard with firm keys was a "drastic departure" from previous Retina MacBook keyboards. It further noted that resting palms may brush the trackpad occasionally, causing inadvertent cursor jumps onscreen as the notebook interprets this as input, without one's hands or wrists actually resting on it. Bandwidth increased; the flash storage was about 40 percent faster. Engadget praised the thinner, lighter design; improved display and audio; and increased speed of the graphics and flash storage, but criticized the lack of ports and the price. Wired praised the display, calling it "the best laptop display I've ever seen", as well as praising the Touch Bar, though it criticized the need of adapters for many common connectors. Likewise, The Verge concluded that "using [the new MacBook] is alienating to anyone living in the present. I agree with Apple's vision of the future. I'm just not buying it today." Engadget voiced their concerns that "by doing things like removing full-sized USB ports, the memory card reader and even the Function row, Apple seems to have forgotten how many of us actually work". Heavy keyboard users criticized the Touch Bar, noting that command-line tools like Vim rely on keyboard usage, and the Touch Bar does not provide the tactile feedback necessary for "blind" usage of Function keys. Miriam Nielsen from The Verge said: "When I tried to intentionally use the Touch Bar, I felt like a kid learning how to type again. I had to keep looking down at the bar instead of looking at the images I was actually trying to edit." She also said that after learning the Touch Bar one cannot work as efficiently on any other computer. Developers have their share of headaches because they cannot rely on the Touch Bar being present on every machine that runs their software. Even if Apple makes the Touch Bar an integral part of macOS, it will take "many years" for it to become ubiquitous, in the meantime, anything in the Bar needs to be available through another part of the interface. Also criticized were non-compatibility between Thunderbolt 2 and 3 devices. Some found unpleasant the fan whine on the 15" model, where the two integrated fans run all the time by default, thanks to the coprocessor powering the Touch Bar and higher TDP of the stronger CPU models. In 2016 and 2017, the Touch Bar caused concern among American state bars that the predictive text could be used to cheat on bar exams. The responses varied state by state: New York State Bar Association banned the use of the MacBook Pro on bar exams; while North Carolina Bar Association allowed students to take the state bar exam with the computer once a proctor verified that the predictive text feature had been disabled. Technical specifications Magic Keyboard revision Apple unveiled the fifth revision of the fourth generation MacBook Pro in 2020, the 16-inch MacBook Pro on November 13, 2019, replacing the 15-inch model. Similar in size to the 15-inch model, it has a larger 16-inch 3072x1920 Retina display set in a narrower bezel, the largest MacBook screen since the 17-inch unibody MacBook Pro that was discontinued in 2012. It has a physical Escape key, a Touch Bar, and a now-separate sapphire-glass-covered Touch ID sensor at the right end of the Touch Bar that doubles as a power button. It uses a scissor mechanism keyboard almost identical to Apple's wireless Magic Keyboard, providing more travel than the previous revision's "Butterfly" keyboard. Like its predecessor, the 16-inch MacBook Pro has four combined Thunderbolt 3 ports that support USB-C 3.1 Gen 2 and dual DisplayPort 1.4 signals, providing 6016×3384 output to run the Pro Display XDR at full resolution. Any port can be used for charging, it includes a 96W USB-C power adapter. At launch only the included adapter and the Pro Display XDR provide full host power. Peripherals that delivered 87W for the 15-inch model, such as LG Ultrafine displays, are recommended to be used with a separate power supply. It also has a 3.5mm headphone jack. It uses the same Coffee Lake CPUs as the 2019 15-inch model. Purchasers can choose between AMD Radeon Pro 5300M or 5500M GPUs with up to 8GB of GDDR6 memory (or from June 2020 onwards, a 5600M GPU with 8GB of HBM2 memory), up to 64GB of 2667MHz DDR4 memory, and up to 8 TB of SSD storage. It includes better speakers, a better three-microphone array, and a 100 Wh battery, the largest that can be easily carried onto a commercial airliner under U.S. Transportation Security Administration rules. On May 4, 2020, Apple announced an updated 13-inch model with the Magic Keyboard. The four Thunderbolt port version comes with Ice Lake processors, updated graphics, up to 32GB of memory and 4TB of storage, and supports 6K output to run the Pro Display XDR. The two Thunderbolt port version has the same Coffee Lake processors, graphics, and maximum storage and memory as the 2019 two Thunderbolt port models. The 2020 13-inch models also gain 0.02 inches (0.6mm) in thickness over the 2019 models. Reception Reception to the 16-inch MacBook Pro was generally positive. LaptopMag called the keyboard "much-improved". The Verge praised the new keyboard, microphones, and speakers, but criticized the lack of peripherals such as an SD card slot. 9to5Mac criticized the use of a 720p webcam and older 802.11ac Wi-Fi standard, noting that Apple's iPhone 11 family included a 4K front-facing camera and faster Wi-Fi 6. MacWorld'' also noted the lack of Face ID. Another review noted that the 2020 two Thunderbolt port 13-inch model is unable to run Apple's Pro Display XDR at full resolution, while the lower-priced 2020 MacBook Air can. There are numerous reports of cracked screens caused by closing the unit with a third-party physical webcam cover due to reduced clearance compared to previous models. Technical specifications Software and operating systems The macOS operating system has been pre-installed on all MacBook Pro computers since release, starting with version 10.4.4 (Tiger). Along with OS X, iLife has also shipped with all systems, beginning with iLife '06. The Intel-based MacBook Pro comes with the successor to BIOS, Extensible Firmware Interface (EFI) 1.1. EFI handles booting differently from BIOS-based computers, but provides backwards compatibility, allowing dual- and triple-boot configurations. In addition to OS X, the Microsoft Windows operating system is installable on Intel x86-based Apple computers. Officially, this is limited to 32-bit versions of Windows XP, Vista, and 7, and 64-bit versions of Windows Vista, 7, 8, 8.1, and 10 with the necessary hardware drivers included with the Boot Camp software. Other x86 operating systems such as Linux are also unofficially supported. This is made possible by the presence of the Intel architecture as provided by the CPU and the BIOS emulation Apple has provided on top of EFI. macOS Big Sur, the seventeenth major release of macOS, will work with Wi-Fi and graphics acceleration on unsupported MacBook Pro computers with a compatible patch utility. See also Comparison of Macintosh models MacBook (12-inch) MacBook Air Notes References External links – official site Computer-related introductions in 2006 MacBook X86 Macintosh computers
865300
https://en.wikipedia.org/wiki/IBM%20305%20RAMAC
IBM 305 RAMAC
The IBM 305 RAMAC was the first commercial computer that used a moving-head hard disk drive (magnetic disk storage) for secondary storage. The system was publicly announced on September 14, 1956, with test units already installed at the U.S. Navy and at private corporations. RAMAC stood for "Random Access Method of Accounting and Control", as its design was motivated by the need for real-time accounting in business. History The first RAMAC to be used in the U.S. auto industry was installed at Chrysler's MOPAR Division in 1957. It replaced a huge tub file which was part of MOPAR's parts inventory control and order processing system. During the 1960 Olympic Winter Games in Squaw Valley (USA), IBM provided the first electronic data processing systems for the Games. The system featured an IBM RAMAC 305 computer, punched card data collection, and a central printing facility. More than 1,000 systems were built. Production ended in 1961; the RAMAC computer became obsolete in 1962 when the IBM 1405 Disk Storage Unit for the IBM 1401 was introduced, and the 305 was withdrawn in 1969. Overview The first hard disk unit was shipped September 13, 1956. The additional components of the computer were a card punch, a central processing unit, a power supply unit, an operator's console/card reader unit, and a printer. There was also a manual inquiry station that allowed direct access to stored records. IBM touted the system as being able to store the equivalent of 64,000 punched cards. The 305 was one of the last vacuum tube computers that IBM built. It weighed over a ton. The IBM 350 disk system stored 5 million alphanumeric characters recorded as six data bits, one parity bit and one space bit for eight bits recorded per character. It had fifty disks. Two independent access arms moved up and down to select a disk, and in and out to select a recording track, all under servo control. Average time to locate a single record was 600 milliseconds. Several improved models were added in the 1950s. The IBM RAMAC 305 system with 350 disk storage leased for US$3,200 () per month. The original 305 RAMAC computer system could be housed in a room of about 9 m (30 ft) by 15 m (50 ft); the 350 disk storage unit measured around . Currie Munce, research vice president for Hitachi Global Storage Technologies (which has acquired IBM's hard disk drive business), stated in a Wall Street Journal interview that the RAMAC unit weighed over a ton, had to be moved around with forklifts, and was delivered via large cargo airplanes. According to Munce, the storage capacity of the drive could have been increased beyond five megabytes, but IBM's marketing department at that time was against a larger capacity drive, because they did not know how to sell a product with more storage. Programming the 305 involved not only writing machine language instructions to be stored on the drum memory, but also almost every unit in the system (including the computer itself) could be programmed by inserting wire jumpers into a plugboard control panel. Architecture System architecture was documented in the 305 RAMAC Manual of Operation. The 305 was a character-oriented variable "word" length decimal (BCD) computer with a drum memory rotating at 6000 RPM that held 3200 alphanumeric characters. A core memory buffer of 100 characters was used for temporary storage during data transfers. Each character was six bits plus one odd parity bit ("R") composed of two zone bits ("X" and "O") and remaining four binary bits for the value of the digit in the following format: X O 8 4 2 1 R Instructions could only be stored on 20 tracks of the drum memory and were fixed length (10 characters), in the following format: T1 A1 B1 T2 A2 B2 M N P Q Fixed-point data "words" could be any size from one decimal digit up to 100 decimal digits, with the X bit of the least significant digit storing the sign (signed magnitude). Data records could be any size from one character up to 100 characters. Drum memory The drum memory was organized into 32 tracks of 100 characters each. The color code of this table is: Yellow – Storage Blue – Arithmetic Green – Input/output Red – Special function L and M select the same track, containing ten 10-character "Accumulators". As a destination L specifies addition, M specifies subtraction. (Numbers in these accumulators were stored in ten's complement form, with the X bit of the most significant digit storing the sign. The sign of each accumulator was also held in a relay. However the 305 automatically converted between its standard signed magnitude format and this format without the need for special programming.) J, R, and - do not select tracks on the drum, they specify other sources and destinations. Jumps The 305's instruction set does not include any jumps, instead these are programmed on the control panel: Unconditional jump – the program exit code (P field) specifies a Program exit hub on the control panel, which has a wire plugged into it and, via distributors, to Program entry hubs specifying the first, second and third address digit of the instruction to jump to. Conditional jump – the program exit code (P field) specifies a Program exit hub on the control panel, which has a wire plugged into it and the appropriate Condition selector common hub to be tested, the corresponding two Condition selector output hubs have wires plugged into them and the Program entry hubs specifying the instructions to jump to or the Program advance hub to continue in sequence. Complicated conditions involving many Condition selectors could be wired to execute in a single instruction (e.g., Testing the sign and zero states of multiple accumulators), with one of several Program entry hubs activated. Multi-way jump – the destination track (T2 field) is set to - and the appropriate Character selector hubs on the control panel have wires plugged into them and the Program entry hubs specifying the instructions to jump to or the Program advance hub to continue in sequence. Timing All timing signals for the 305 were derived from a factory recorded clock track on the drum. The clock track contained 816 pulses 12 μs apart with a 208 μs gap for sync. Reading or writing a character took 96 μs. The 305's typical instruction took three revolutions of the drum (30 ms): one (I phase) to fetch the instruction, one (R phase) to read the source operand and copy it to the core buffer, and one (W phase) to write the destination operand from the core buffer. If the P field (Program exit code) was not blank, then two (D phase and P phase) additional revolutions of the drum (20 ms) were added to the execution time to allow relays to be picked. The Improved Processing Speed option could be installed that allowed the three instruction phases (IRW) to immediately follow each other instead of waiting for the next revolution to start; with this option and well optimized code and operand placement a typical instruction could execute in as little as one revolution of the drum (10 ms). Certain instructions though took far longer than the typical 30 ms to 50 ms. For example, multiply took six to nineteen revolutions of the drum (60 ms to 190 ms) and divide (an option) took ten to thirty seven revolutions of the drum (100 ms to 370 ms). Input/Output instructions could interlock the processor for as many revolutions of the drum as needed by the hardware. Hardware implementation The logic circuitry of the 305 was built of one- and two-tube pluggable units and relays. Related peripheral units A basic system was composed of the following units: IBM 305 – Processing unit, the magnetic process drum, magnetic core register and electronic logical and arithmetic circuits IBM 350 – Disk storage unit IBM 370 – Printer IBM 323 – Card punch IBM 380 – Console, the card reader and IBM Electric typewriter model B1 IBM 340 – Power supply See also List of vacuum tube computers History of hard disk drives References External links IBM 305 RAMAC Data Processing System IBM Archives on the 305 IBM 350 RAMAC site originally prepared under the supervision of the Storage Special Interest Group of the Computer History Museum Youtube video 305 Computer-related introductions in 1956
63673072
https://en.wikipedia.org/wiki/Tandy%203000
Tandy 3000
The Tandy 3000 is a personal computer introduced by Radio Shack in 1986 based on the 16-bit 8 MHz Intel 80286 microprocessor. Description The Tandy 3000 is functionally a clone of the IBM PC-AT, the first PC by a major manufacturer using the fully 16-bit Intel 286 processor. As such, it departed from Tandy's two previous PC workalikes (the Tandy 2000 in 1983 and the Tandy 1000 in 1985) in that it was built without proprietary technology. The motherboard contains no built-in circuitry for its disk controller or video display. Owners could outfit the computer, and upgrade it, with standard PC components sold by Tandy or available from third-party suppliers. Since the hardware is industry-standard throughout, there were no compatibility issues such as there were with the previous models 2000 and 1000. More accurately, any compatibility troubles that might arise were no fault of the computer, but rather, any third-party hardware installed or with the AT architecture upon which the computer was engineered. The operating system was an extra-cost item; the purchaser could choose MS-DOS 3.2 or Xenix V. Xenix and the extra memory it demanded was expensive but permitted up to six remote terminals to run programs on a single Tandy 3000 simultaneously. Microsoft's BASIC interpreter, bundled with Tandy's Deskmate productivity suite, was offered at extra cost. Digital Research's CP/M-86 was an option available from other software vendors. Later, others available for generic AT clones such as the Tandy 3000 included IBM's PC DOS, Digital Research's DR-DOS and GEM, and 16-bit versions of Microsoft's Windows (up to version 3.x). Still later IBM's graphical multitasking OS/2 was an option for machines equipped with enough memory and capable graphics display hardware. Base memory was 512 KB, expandable to 640 KB on the motherboard. RAM was expandable to a maximum of 12 MB using cards in the expansion slots. The Tandy 3000 has ten expansion slots: seven 16-bit AT compatible, two 8-bit XT compatible, and one half-size XT slot dedicated to the included serial/parallel card. After the disk controller card and the video card are installed six slots are available for expansion, which is one more than the PC-AT offers. On the other hand, the Tandy computer lacks the AT's locking keyswitch for operational security. There are three drive bays available; two are accessible through the front panel, so two may be floppy drives or other replaceable media like a PCMCIA card reader. A single 1.2 MB 5.25 inch floppy disk drive was included in the base configuration. This drive was first introduced by IBM for the PC-AT, and was unique to it. Like IBM's later PS/2 series and other newer computers in the industry, Tandy's follow-on computers would discontinue using this drive and switch to the 3.5 inch format. From Tandy two sizes of hard drive were available: 20 MB and 35 MB. Being a generic AT clone wearing the Tandy badge, any third-party drives compatible with the AT could be installed. In later years drives with capacities far greater than 35 MB became available for installation, though the MS-DOS 3.2 most often purchased with the computer could not manage so much storage without partitioning. Two hard drives could be installed in the cabinet, provided the controller card could handle two. There is a socket on the motherboard for the 287 math co-processor. The Tandy 3000 has a smaller desk footprint than the AT; it occupies 342 square inches (19 by 18 inches) which is less than the AT's footprint of 367.2 square inches (21.25 by 17.28 inches). Two years after its introduction, the Tandy 3000 was improved with a 286 processor that ran at a clock rate of 12 MHz, which could be switched to 6 MHz. The motherboard was revised to permit up to 16 MB of RAM in the expansion slots, the most memory the 286 chip could address. Installed RAM increased to 640 KB. The AT's locking keyswitch was built in. Tandy 3000 HL In 1987 Tandy introduced the 3000 HL, which was aimed at the budget market. This computer is functionally similar with less expandability (only three 16-bit AT slots and four 8-bit XT slots, which limit maximum RAM to 4 megabytes). It was limited to 8 MHz clock speed. It still has three drive bays and its new cabinet has a smaller desk footprint: 17 by 15.5 inches. Tandy 3000 NL In 1989 Tandy updated the 3000 HL with several improvements: clock speed of 10 MHz and a 3.5 inch high density (1.44 MB) floppy disk drive. The all-new cabinet has two 3.5 inch and two 5.25 inch drive bays. It has the same expansion slots as the 3000 HL (three AT and four XT slots) plus one high-speed 10 MHz slot dedicated to memory expansion. Maximum RAM that can be installed in the slots is now sixteen megabytes. The computer has a locking keyswitch. In 1990 Tandy updated the 3000 NL's motherboard with built-in VGA circuitry, obviating the need to install a video card in an expansioni slot. Tandy 2500 XL From 1991 the 3000 line was dropped and replaced with Tandy's last PC-AT equivalent, the 2500 XL. This was housed in a new cabinet measuring 15 by 15.5 inches, with two 3.5 inch and one 5.25 inch drive bays. One high density 3.5 inch 1.44 MB floppy drive comes installed. There are three 16-bit expansion slots, all of which run at the full CPU clock rate of 10 MHz. One megabyte of RAM comes installed, and using expansion slots RAM can go to 16 megabytes (the maximum addressable by the 286). The motherboard has a built-in SmartDrive hard drive controller. The parallel port is now bidirectional, and there is a mouse port. By this time the new Intel 80386SX processor, with its 16-bit external data bus, 24-bit address bus and 32-bit processor core, enabled PC makers to offer performance superior to the 286/PC-AT at more affordable prices. The full 32-bit bus of the 386DX required a more expensive circuit board and peripheral chips. Accordingly, starting in 1991, Tandy followed the industry trend of substituting computers using it for products based on the 286/PC-AT technology. Consequently the new 4016 and 4020 SX computers replaced the 2500 XL, leaving only the 1000 TL (and later the 1000 RLX) line of desktops to continue using the 286 processor, which had 8-bit XT expansion slots, not 16-bit AT compatible slots. Further, the 1000s could not run the graphical multitasking OS/2 operating system, as could the 3000 and 2500 XL. In 1992 Tandy offered an upgraded 2500 XL with its 286 processor clocked at 16 MHz and a Super VGA (maximum pixel resolution of 1024 by 768) video card built into the motherboard. The one megabyte standard RAM is expandable to five megabytes without using an expansion slot. See also Tandy 1000 Tandy 2000 Tandy 6000 TRS-80 Model II List of TRS-80 and Tandy-branded computers Note References External links About Tandy 3000 on Orlando Sentinel The web is now 25 years old and the computer equipment we use today is exponentially cheaper and faster than in 1989 Tandy advertisement for Tandy 3000 Home computers IBM PC compatibles RadioShack
17455275
https://en.wikipedia.org/wiki/Display%20aspect%20ratio
Display aspect ratio
The aspect ratio of a display device is the proportional relationship between the width and the height of the display. It is expressed as two numbers separated by a colon (x:y), where x corresponds to the width and y to the height. Common aspect ratios for displays, past and present, include 5:4, 4:3, 16:10 and 16:9. Computer displays As of 2016, most computer monitors use widescreen displays with an aspect ratio of 16:9, although some portable PCs use narrower aspect ratios like 3:2 and 16:10 while some high-end desktop monitors have adopted ultrawide displays. The following table summarises the different aspect ratios that have been used in computer displays: † The resolution doesn't match the aspect ratio exactly, but is commonly marketed or described as such. History 4:3, 5:4 and 16:10 Until about 2003, most computer monitors used an aspect ratio of 4:3, and in some cases 5:4. For cathode ray tubes (CRT)s 4:3 was most common even in resolutions where this meant the pixels would not be square (e.g. 320×200 or 1280×1024 on a 4:3 display). Between 2003 and 2006, monitors with 16:10 aspect ratio became commonly available, first in laptops and later also in standalone computer monitors. Reasons for this transition was productive uses for such monitors, i.e. besides widescreen movie viewing and computer game play, are the word processor display of two standard A4 or letter pages side by side, as well as CAD displays of large-size drawings and CAD application menus at the same time. 16:10 became the most common sold aspect ratio for widescreen computer monitors until 2008. 16:9 In 2008, the computer industry started to move from 4:3 and 16:10 to 16:9 as the standard aspect ratio for monitors and laptops. A 2008 report by DisplaySearch cited a number of reasons for this shift, including the ability for PC and monitor manufacturers to expand their product ranges by offering products with wider screens and higher resolutions, helping consumers to more easily adopt such products and "stimulating the growth of the notebook PC and LCD monitor market". By 2010, virtually all computer monitor and laptop manufacturers had also moved to the 16:9 aspect ratio, and the availability of 16:10 aspect ratio in mass market had become very limited. In 2011, non-widescreen displays with 4:3 aspect ratios still were being manufactured, but in small quantities. The reasons for this according to Bennie Budler, product manager of IT products at Samsung South Africa was that the "demand for the old 'Square monitors' has decreased rapidly over the last couple of years". He also predicted that "by the end of 2011, production on all 4:3 or similar panels will be halted due to a lack of demand." In 2012, 1920×1080 was the most commonly used resolution among Steam users. At the same time, the most common resolution globally was 1366×768, overtaking the previous leader 1024×768. In 2021, the 2K resolution of 1920×1080 was used by two third of the Steam users for the primary display with 1366×768 and 2560×1440 both at about eight percent taking the majority of the remaining resolutions. 3:2 3:2 displays first appeared in laptop computers in 2001 with the PowerBook G4 line, but didn't enter the mainstream until the 2010s with the Chromebook Pixel and 2-in-1 PCs like Microsoft's Surface line. As of 2018, a number of manufacturers are either producing or planning to produce portable PCs with 3:2 displays. 21:9 Since 2014, a number of high-end desktop monitors have been released that use ultrawide displays with aspect ratios that roughly match the various anamorphic formats used in film, but are commonly marketed as 21:9. Resolutions for such displays include 2560×1080 (64:27), 3440×1440 (43:18) and 3840×1600 (12:5). 32:9 In 2017, Samsung released a curved gaming display with an aspect ratio of 32:9 and resolution of 3840×1080. 256:135 Since 2011, several monitors complying with the Digital Cinema Initiatives 4K standard have been produced; this standard specifies a resolution of 4096×2160, giving an aspect ratio of ≈1.896:1. 1:1 A 1:1 aspect ratio, results in a square display. One of the available monitors for desktop use of this format is Eizo EV2730Q (27", 1920 × 1920 Pixels, from 2015), however such monitors are also often found in air traffic control displays (connected using standard computer cabling, like DVI or DisplayPort) and on aircraft as part of avionic equipment (often connected directly using LVDS, SPI interfaces or other specialized means). This 1920×1920 display can also be used as the centerpiece of a three-monitor array with one WUXGA set in vertical position on each side, resulting in 4320×1920 (a ratio of 9:4) - and no distortion with the Eizo 27" 1:1 if the side displays are 22". Suitability for software and content Games From 2005 to 2013 most video games were mainly made for the 16:9 aspect ratio and 16:9 computer displays therefore offer the best compatibility. 16:9 video games are letterboxed on a 16:10 or 4:3 display or have reduced field of view. As of 2013, many games are adopting support for 21:9 ultrawide resolutions, which can give a gameplay advantage due to increased field of view, although this is not always the case. 4:3 monitors have the best compatibility with older games released prior to 2005 when that aspect ratio was the mainstream standard for computer displays. Video As of 2017, the most common aspect ratio for TV broadcasts is 16:9, whereas movies are generally made in the wider 21:9 aspect ratio. Most modern TVs are 16:9, which causes letterboxing when viewing 21:9 content, and pillarboxing when viewing 4:3 content such as older films or TV broadcasts, unless the content is cropped or stretched to fill the entire display. Windows Microsoft recommends a 16:9 display for tablet computers running Windows 8. Productivity applications Microsoft recommends a 16:9 display for Office 2013. For viewing documents in A4 paper size (which has a 1.41:1 aspect ratio), whether in portrait mode or two side-by-side in landscape mode, 4:3 or 16:10 fits best. For photographs in the standard 135 film and print size (with a 3:2 aspect ratio), 16:10 fits best; for photographs taken with consumer-level digital cameras, 4:3 fits perfectly. Diagonal and area The size of a computer monitor is given as the diagonal measurement of its display area, usually in inches. Wider aspect ratios result in smaller overall area, given the same diagonal. Smartphones Until 2010, smartphones used different aspect ratios, including 3:2 and 5:3. Since then, most smartphone manufacturers have switched to using 16:9 widescreen displays, driven at least partly by the growing popularity of HD video using the same aspect ratio. Since 2017, a number of smartphones have been released using 18:9 or even wider aspect ratios (such as 18.5:9 or 19.5:9); such displays are expected to appear on increasingly more phones. Reasons for this trend include the ability for manufacturers to use a nominally larger display without increasing the width of the phone, being able to accommodate the on-screen navigation buttons without reducing usable app area, more area available for split-screen apps in portrait orientation, as well as the 18:9 ratio being well-suited for VR applications and the proposed Univisium film format. On the other hand, the disadvantages of taller 18:9 aspect ratio phones with some phones even going up to 20:9 or 21:9 is reduced one-handed reachability, being less convenient to carry around in the pocket as they stick out and reduced overall screen surface area. TVs Most televisions were built with an aspect ratio of 4:3 until the early 2010s, when widescreen TVs with 16:9 displays became the standard. This aspect ratio was chosen as the geometric mean between 4:3 and 2.35:1, an average of the various aspect ratios used in film. While 16:9 is well-suited for modern HDTV broadcasts, older 4:3 video has to be either padded with bars on the left and right side (pillarboxed), cropped or stretched, while movies shot with wider aspect ratios are usually letterboxed, with black bars at the top and bottom. Since turn of the 21st century, many music videos began shooting on widescreen aspect ratio. See also 14:9 aspect ratio Computer monitor Display resolution Field of view in video games Graphics display resolution Ultrawide formats References Display technology Engineering ratios
23189
https://en.wikipedia.org/wiki/Shuffling
Shuffling
Shuffling is a procedure used to randomize a deck of playing cards to provide an element of chance in card games. Shuffling is often followed by a cut, to help ensure that the shuffler has not manipulated the outcome. Techniques Overhand One of the easiest shuffles to accomplish after a little practice is the overhand shuffle. Johan Jonasson wrote, "The overhand shuffle... is the shuffling technique where you gradually transfer the deck from, say, your right hand to your left hand by sliding off small packets from the top of the deck with your thumb." In detail as normally performed, with the pack initially held in the left hand (say), most of the cards are grasped as a group from the bottom of the pack between the thumb and fingers of the right hand and lifted clear of the small group that remains in the left hand. Small packets are then released from the right hand a packet at a time so that they drop on the top of the pack accumulating in the left hand. The process is repeated several times. The randomness of the whole shuffle is increased by the number of small packets in each shuffle and the number of repeat shuffles performed. The overhand shuffle offers sufficient opportunity for sleight of hand techniques to be used to affect the ordering of cards, creating a stacked deck. The most common way that players cheat with the overhand shuffle is by having a card at the top or bottom of the pack that they require, and then slipping it to the bottom at the start of a shuffle (if it was on top to start), or leaving it as the last card in a shuffle and just dropping it on top (if it was originally on the bottom of the deck). Riffle A common shuffling technique is called the riffle, or dovetail shuffle or leafing the cards, in which half of the deck is held in each hand with the thumbs inward, then cards are released by the thumbs so that they fall to the table interleaved. Many also lift the cards up after a riffle, forming what is called a bridge which puts the cards back into place; it can also be done by placing the halves flat on the table with their rear corners touching, then lifting the back edges with the thumbs while pushing the halves together. While this method is more difficult, it is often used in casinos because it minimizes the risk of exposing cards during the shuffle. There are two types of perfect riffle shuffles: if the top card moves to be second from the top then it is an in shuffle, otherwise it is known as an out shuffle (which preserves both the top and bottom cards). The Gilbert–Shannon–Reeds model provides a mathematical model of the random outcomes of riffling that has been shown experimentally to be a good fit to human shuffling and that forms the basis for a recommendation that card decks be riffled seven times in order to randomize them thoroughly. Later, mathematicians Lloyd M. Trefethen and Lloyd N. Trefethen authored a paper using a tweaked version of the Gilbert–Shannon–Reeds model showing that the minimum number of riffles for total randomization could also be six, if the method of defining randomness is changed. Hindu Also known as the "Indian", "Kattar", "Kenchi" (Hindi for scissor) or "Kutti Shuffle". The deck is held face down, with the middle finger on one long edge and the thumb on the other on the bottom half of the deck. The other hand draws off a packet from the top of the deck. This packet is allowed to drop into the palm. The maneuver is repeated over and over, with newly drawn packets dropping onto previous ones, until the deck is all in the second hand. Indian shuffle differs from stripping in that all the action is in the hand taking the cards, whereas in stripping, the action is performed by the hand with the original deck, giving the cards to the resulting pile. This is the most common shuffling technique in Asia and other parts of the world, while the overhand shuffle is primarily used in Western countries. Pile Cards are simply dealt out into a number of piles, then the piles are stacked on top of each other. Though this is deterministic and does not randomize the cards at all, it ensures that cards that were next to each other are now separated. Some variations on the pile shuffle attempt to make it slightly random by dealing to the piles in a random order each circuit. Corgi Also known as the Chemmy, Irish, wash, scramble, beginner shuffle, smooshing, schwirsheling, or washing the cards, this involves simply spreading the cards out face down, and sliding them around and over each other with one's hands. Then the cards are moved into one pile so that they begin to intertwine and are then arranged back into a stack. This method is useful for beginners, but the shuffle requires a large surface for spreading out the cards. Statistically random shuffling is achieved after approximately one minute of smoothing. Smooshing has been largely popularized by Simon Hofman. Mongean The Mongean shuffle, or Monge's shuffle, is performed as follows (by a right-handed person): Start with the unshuffled deck in the left hand and transfer the top card to the right. Then repeatedly take the top card from the left hand and transfer it to the right, putting the second card at the top of the new deck, the third at the bottom, the fourth at the top, the fifth at the bottom, etc. The result, if one started with cards numbered consecutively , would be a deck with the cards in the following order: . For a deck of given size, the number of Mongean shuffles that it takes to return a deck to starting position, is known . Twelve perfect Mongean shuffles restore a 52-card deck. Faro Weaving is the procedure of pushing the ends of two halves of a deck against each other in such a way that they naturally intertwine. Sometimes the deck is split into equal halves of 26 cards which are then pushed together in a certain way so as to make them perfectly interweave. This is known as a Faro Shuffle. The faro shuffle is performed by cutting the deck into two, preferably equal, packs in both hands as follows (right-handed): The cards are held from above in the right and from below in the left hand. Separation of the deck is done simply lifting up half the cards with the right hand thumb slightly and pushing the left hand's packet forward away from the right hand. The two packets are often crossed and slammed into each other as to align them. They are then pushed together by the short sides and bent (either up or down). The cards then alternately fall into each other, much like a zipper. A flourish can be added by springing the packets together by applying pressure and bending them from above, as called the bridge finish. The faro is a controlled shuffle which does not randomize a deck when performed properly. A perfect faro shuffle, where the cards are perfectly alternated, is considered one of the most difficult sleights by card magicians, simply because it requires the shuffler to be able to cut the deck into two equal packets and apply just the right amount of pressure when pushing the cards into each other. Performing eight perfect faro shuffles in a row restores the order of the deck to the original order only if there are 52 cards in the deck and if the original top and bottom cards remain in their positions (1st and 52nd) during the eight shuffles. If the top and bottom cards are weaved in during each shuffle, it takes 52 shuffles to return the deck back into original order (or 26 shuffles to reverse the order). Mexican spiral The Mexican spiral shuffle is performed by cyclic actions of moving the top card onto the table, then the new top card under the deck, the next onto the table, next under the deck, and so on until the last card is dealt onto the table. It takes quite a long time, compared with riffle or overhand shuffles, but allows other players to fully control cards which are on the table. The Mexican spiral shuffle was popular at the end of the 19th century in some areas of Mexico as a protection from gamblers and con men arriving from the United States. Faking Magicians, sleight-of-hand artists, and card cheats employ various methods of shuffling whereby the deck appears to have been shuffled fairly, when in reality one or more cards (up to and including the entire deck) stays in the same position. It is also possible, though generally considered very difficult, to "stack the deck" (place cards into a desirable order) by means of one or more riffle shuffles; this is called "riffle stacking". Both performance magicians and card sharps regard the Zarrow shuffle and the Push-Through-False-Shuffle as particularly effective examples of the false shuffle. In these shuffles, the entire deck remains in its original order, although spectators think they see an honest riffle shuffle. Machines Casinos often equip their tables with shuffling machines instead of having croupiers shuffle the cards, as it gives the casino a few advantages, including an increased complexity to the shuffle and therefore an increased difficulty for players to make predictions, even if they are collaborating with croupiers. The shuffling machines are carefully designed to avoid biasing the shuffle and are typically computer-controlled. Shuffling machines also save time that would otherwise be wasted on manual shuffling, thereby increasing the profitability of the table. These machines are also used to lessen repetitive-motion-stress injuries to a dealer. Players with superstitions often regard with suspicion any electronic equipment, so casinos sometimes still have the croupiers perform the shuffling at tables that typically attract those crowds (e.g., baccarat tables). Randomization There are exactly 52 factorial (expressed in shorthand as 52!) possible orderings of the cards in a 52-card deck. In other words, there are 52 × 51 × 50 × 49 × ··· × 4 × 3 × 2 × 1 possible combinations of card sequence. This is approximately (80,658vigintillion) possible orderings, or specifically 80,658,175,170,943,878,571,660,636,856,403,766,975,289,505,440,883,277,824,000,000,000,000. The magnitude of this number means that it is exceedingly improbable that two randomly selected, truly randomized decks will be the same. However, while the exact sequence of all cards in a randomized deck is unpredictable, it may be possible to make some probabilistic predictions about a deck that is not sufficiently randomized. Sufficiency The number of shuffles that are sufficient for a "good" level of randomness depends on the type of shuffle and the measure of "good enough randomness", which in turn depends on the game in question. For most games, four to seven riffle shuffles are sufficient: for unsuited games such as blackjack, four riffle shuffles are sufficient, while for suited games, seven riffle shuffles are necessary. There are some games, however, for which even seven riffle shuffles are insufficient. In practice the number of shuffles required depends both on the quality of the shuffle and how significant non-randomness is, particularly how good the people playing are at noticing and using non-randomness. Two to four shuffles is good enough for casual play. But in club play, good bridge players take advantage of non-randomness after four shuffles, and top blackjack players supposedly track aces through the deck; this is known as "ace tracking", or more generally, as "shuffle tracking". Research Following early research at Bell Labs, which was abandoned in 1955, the question of how many shuffles was required remained open until 1990, when it was convincingly solved as seven shuffles, as elaborated below. Some results preceded this, and refinements have continued since. A leading figure in the mathematics of shuffling is mathematician and magician Persi Diaconis, who began studying the question around 1970, and has authored many papers in the 1980s, 1990s, and 2000s on the subject with numerous co-authors. Most famous is , co-authored with mathematician Dave Bayer, which analyzed the Gilbert–Shannon–Reeds model of random riffle shuffling and concluded that the deck did not start to become random until five good riffle shuffles, and was truly random after seven, in the precise sense of variation distance described in Markov chain mixing time; of course, you would need more shuffles if your shuffling technique is poor. Recently, the work of Trefethen et al. has questioned some of Diaconis' results, concluding that six shuffles are enough. The difference hinges on how each measured the randomness of the deck. Diaconis used a very sensitive test of randomness, and therefore needed to shuffle more. Even more sensitive measures exist, and the question of what measure is best for specific card games is still open. Diaconis released a response indicating that you only need four shuffles for un-suited games such as blackjack. On the other hand, variation distance may be too forgiving a measure and seven riffle shuffles may be many too few. For example, seven shuffles of a new deck leaves an 81% probability of winning New Age Solitaire where the probability is 50% with a uniform random deck. One sensitive test for randomness uses a standard deck without the jokers divided into suits with two suits in ascending order from ace to king, and the other two suits in reverse. (Many decks already come ordered this way when new.) After shuffling, the measure of randomness is the number of rising sequences that are left in each suit. Algorithms If a computer has access to purely random numbers, it is capable of generating a "perfect shuffle", a random permutation of the cards; beware that this terminology (an algorithm that perfectly randomizes the deck) differs from "a perfectly executed single shuffle", notably a perfectly interleaving faro shuffle. The Fisher–Yates shuffle, popularized by Donald Knuth, is simple (a few lines of code) and efficient (O(n) on an n-card deck, assuming constant time for fundamental steps) algorithm for doing this. Shuffling can be seen as the opposite of sorting. There are other, less-desirable algorithms in common use. For example, one can assign a random number to each card, and then sort the cards in order of their random numbers. This will generate a random permutation, unless any of the random numbers generated are the same as any others (i.e. pairs, triplets etc.). This can be eliminated either by adjusting one of the pair's values randomly up or down by a small amount, or reduced to an arbitrarily low probability by choosing a sufficiently wide range of random number choices. If using efficient sorting such as mergesort or heapsort this is an O(n log n) average and worst-case algorithm. Online gambling These issues are of considerable commercial importance in online gambling, where the randomness of the shuffling of packs of simulated cards for online card games is crucial. For this reason, many online gambling sites provide descriptions of their shuffling algorithms and the sources of randomness used to drive these algorithms, with some gambling sites also providing auditors' reports of the performance of their systems. See also Card manipulation Mental poker Solitaire (cipher) References Footnotes External links Physical card shuffling: Illustrated guide to several shuffling methods Magician's tool with much shuffling simulation Mathematics of shuffling: Real World Shuffling In Practice Shuffle - MathWorld - Wolfram Research Ivars Peterson's MathTrek: Card Shuffling Shenanigans Real world (historical) application: How We Learned to Cheat at Online Poker: A Study in Software Security Card game terminology Card shuffling
9613511
https://en.wikipedia.org/wiki/Digital%20economy
Digital economy
The digital economy is an economy that is based on digital computing technologies, but is often perceived as conducting business through markets based on the internet and the World Wide Web. It is also known as the Internet Economy, New Economy, or Web Economy. The digital economy is intertwined with the traditional economy, making a clear delineation harder. The digital economy results from billions of everyday online connections among people, businesses, devices, data, & processes. It is based on the interconnectedness of people, organizations, and machines that results from the Internet, mobile technology and the internet of things (IoT). Without the Internet, the digital economy that the global economy runs on would not exist in its current form. The digital economy is backed by the spread of information and communication technologies (ICT) across all business sectors to enhance its productivity. Digital transformation of the economy is altering conventional notions about how businesses are structured, how consumers obtain goods and services, and how states need to adapt to new regulatory challenges. The future of work, especially since the COVID-19 pandemic, is also contributing to the digital economy. More people are now working online, and with the increase of online activity that contributes to the global economy, companies that support the systems of the Internet are more profitable. Definition The Digital Economy also referred to as the New Economy, refers to an Economy in which digital computing technologies are used in Economic Activities. The term Digital Economy was first mentioned in Japan by a Japanese professor and research economist in the midst of Japan's recession of the 1990s. In the West, the term came into use during the early 1990s. For example, many academic papers were published by New York University’s Center for Digital Economy Research. The term was more widely popularized by Don Tapscott's 1995 book, The Digital Economy: Promise and Peril in the Age of Networked Intelligence. This book explained how the Internet would change the way the world did business. According to Thomas Mesenbourg (2001), three main components of the digital economy concept can be identified: E-business infrastructure (hardware, software, telecom, networks, human capital, etc.), E-business (how business is conducted, any process that an organization conducts over computer-mediated networks), E-commerce (transfer of goods, for example when a book is sold online). Bill Imlah states that new applications are blurring these boundaries and adding complexity, for example, social media and Internet search. In the last decade of the 20th century, Nicholas Negroponte (1995) used a metaphor of shifting from processing atoms to processing bits: "The problem is simple. When information is embodied in atoms, there is a need for all sorts of industrial-age means and huge corporations for delivery. But suddenly, when the focus shifts to bits, the traditional big guys are no longer needed. Do-it-yourself publishing on the Internet makes sense. It does not for a paper copy." In this new economy, digital networking and communications infrastructure provide a global platform on which people and organizations devise strategies, interact, communicate, collaborate, and search for information. More recently, "Digital Economy" has been defined as the branch of economics studying the movement of "zero marginal cost intangible goods" over the Internet. Development of the concept The definition of Digital Economy (or similar concepts) is not harmonized across governments, businesses, and international organizations. According to the OECD, the Digital Economy can be defined in three different approaches: Bottom-up approach: characterizing industries’ and firms‘ output or production processes to decide whether they should be included in the Digital Economy, Top-down or trend-based approach: first identifying the key trends driving the digital transformation and then analyzing the extent to which these are reflected in the real economy, Flexible or tiered approach: breaking the Digital Economy into core and non-core components, and thereby finding a compromise between adaptability and the need to arrive at some common ground on the meaning of the term. Bottom-up definition Bottom-up definitions define the Digital Economy as the aggregate of a specific indicator for a set of industries identified as actors in the Digital Economy. Whether an industry is considered an actor depends on the nature of the products (narrow) or the proportion of digital inputs used in production processes (broad). Hence, from a bottom-up and narrow perspective, the Digital Economy is "all industries or activities that directly participate in producing, or crucially reliant on digital inputs." For instance, McKinsey adds up the economic outputs of the ICT sector and e-commerce market in terms of online sales of goods and consumer spending on digital equipment. While this definition is adept at measuring the impact of digitalization on economic growth, it only focuses on the nature of output and offers an incomplete view of the Digital Economy's development. In a bottom-up and broad perspective, the Digital Economy is "all industries using digital inputs as part of their production process". Examples of digital inputs include digital infrastructure, equipment, and software but can include data and digital skills. Top-down definition Top-down definitions identify broad trends at play in the digital transformation and define the Digital Economy as the result of their combined impact on value creation. These include such spillovers as changes in labor market demand and regulations, platform economy, sustainability, and equality. Unlike the bottom-up definition, the top-down definition has units of analysis extending beyond firms, industries, and sectors to include individuals, communities, and societies. While the latter definition is more inclusive, the IMF notes that it is subjective, qualitative, and open-ended, thus limiting meaningful comparative analysis. Flexible definition To reconcile the bottom-up and top-down definitions of the Digital Economy, Bukht and Heeks stated that the Digital Economy consists of all sectors making extensive use of digital technologies (i.e. their existence depends on digital technologies), as opposed to sectors making intensive use of digital technologies (i.e. simply employing digital technologies to increase productivity). Under this definition, the Digital Economy is stratified into three nested tiers: Core: comprising the digital sector and associated core technologies. Examples include hardware manufacturing, software and IT consulting, information services, and telecommunications, Narrow scope: the digital economy comprising digital services and the platform-based economy, Broad scope: the digitalized economy comprising digitalized sectors such as e-Business, e-Commerce, advanced manufacturing, precision agriculture, algorithmic economy, sharing economy, and gig economy. These digitalized sectors phenomenologically give rise to the Fourth Industrial Revolution. Gig Economy Gig work is labor that consists of temporary and flexible jobs usually done over delivery apps and rideshare services such as Grubhub, Uber, Lyft, and Uber Eats. It can be desirable to those who want more flexibility in their schedule and can allow workers to make additional income outside of their traditional jobs. Most gig work supplements workers' traditional jobs. The full size of the gig economy and number of workers is not yet known. Katz and Krueger estimated that only 0.5% of gig workers make most of their income off of platforms like Uber, Lyft, Grubhub, and DoorDash. Since these workers are considered independent contractors, these companies are not responsible for giving its workers benefits packages like it would for regular full-time employees. This has resulted in the formation of unions between gig and platform workers and various reforms within the industry. Blockchain and Tokenized equity-sharing gig economy platforms or applications are being developed to accelerate the gig economy as a full fledged digital economy contributor using new technologies. Information Technology The information technology (IT) sector of the U.S. now makes up about 8.2% of the country's GDP and accounts for twice its share of the GDP as compared to the last decade. 45% of spending on business equipment are investments in IT products and services, which is why companies such as Intel, Microsoft, and Dell have grown from $12 million in 1987 to more than half a billion in 1997. The Framework for Global Electronic Commerce In the U.S. in the 1990's, the Clinton Administration proposed The Framework for Global Electronic Commerce. It contained the promotion of five principles used to guide the U.S. government's actions towards electronic commerce so that the digital economy's growth potential remains high. These five principles include the leadership of the private sector, the government avoiding undue restrictions on e-commerce, limited government involvement, the government's recognition of the Internet's unique qualities, and the facilitation of e-commerce on a global basis. Digital platforms A digital platform operator is an entity or person offering an online communication service to the public based on computer algorithms used to classify content, goods, or services offered online, or the connection of several parties for the sale of goods, the provision of a service, or the exchange or sharing of content, goods, and services. Blockchain Blockchain is used for cryptocurrency by using individual nodes that run a system for cryptocurrency exchanges. Exchanges of cryptocurrency are validated by miners. Miners validity is checked by other nodes for correctness. The use of blockchain technology reduces transaction times and fraud. Companies, such as Tesla, take cryptocurrency as a form of payment but do not accept all. In 2021, Tesla began accepting Bitcoin then several months later stopped accepting due to environmental issues with mining of bitcoin. Tesla stopped accepting Bitcoin due to the environmental implications created by mining cryptocurrency. Cryptocurrency mining takes a substantial amount of electricity, generating a significant carbon footprint. Spread of Information, Communication Technologies (ICT) The widespread adoption of ICT combined with the rapid decline in price and increase in the performance of these technologies, has contributed to the development of new activities in the private and public sectors. These new technologies provide market reach, lower costs, and new opportunities for products and services that were not needed before. This changes the way multinational enterprises (MNE) and startups design their business models. Economic Impact The Digital Economy was estimated to be worth three trillion dollars in 2010. This is about 30% of the S&P 500, six times the U.S.’ annual trade deficit or more than the GDP of the United Kingdom. It is widely accepted that the growth of the digital economy has widespread impact on the whole economy. Various attempts at categorizing the size of the impact on traditional sectors have been made. The Boston Consulting Group discussed "four waves of change sweeping over consumer goods and retail". In 2012, Deloitte ranked six industry sectors as having a "short fuse" and to experience a "big bang" as a result of the digital economy. Telstra, an Australian telecommunications provider, describes how competition will become more global and more intense as a result of the digital economy. In 2016, the Digital Economy represented $11.5 trillion, or 15.5% of global GDP (18.4% of GDP in developed economies and 10 per cent in developing economies on average). It found that the digital economy had grown two and a half times faster than global GDP over the previous 15 years, almost doubling in size since 2000. Most of the value in the digital economy was produced in only a few economies: the United States (35%), China (13%) and Japan (8%). The EU together with Iceland, Liechtenstein and Norway accounted for another 25%. Impact on Retail The digital economy has had a substantial impact on retail sales of consumer product goods. One effect has been the fast proliferation of retailers with no physical presence, such as eBay or Amazon. Additionally, traditional retailers such as Walmart and Macy's have restructured their businesses to adapt to a digital economy. Some retailers, like Forever 21, have declared bankruptcy as a result of their failure to anticipate and adapt to a digital economy. Others, such as Bebe stores have worked with outside vendors to completely convert their business one that is exclusively digital. These vendors, such as IBM and Microsoft, have enabled smaller retailers to compete with large, multi-national established brands. Key features Mobility Mobility of intangibles Both development and exploitation of intangible assets are key feature of the digital economy. This investment in and development of intangibles such as software is a core contributor to value creation and economic growth for companies in the digital economy. In early 2000, companies started substantially increasing the amount of capital allocated to intangibles such as branding, design and, technology rather than in hardware, machinery or property. Mobility of business functions Advancements in information and communication technologies (ICT) have significantly reduced the cost associated with the organization and coordination of complex activities over a long period. Some businesses are increasingly able to manage their global operations on an integrated basis from a central location separate geographically from the locations in which the operations are carried out, and where their suppliers or customers are. Consequently, it has allowed businesses to expand access to remote markets and provide goods and services across borders. Reliance on Data The Digital economy relies on personal data collection. In 1995, the Data Protection directive (Directive 95/46/CE, art.2), defined data as "any information relating to a natural person who can be identified by reference to his identification number or to information which is specific to him". At that time, this regulation emerged in response to the need to integrate the European market. By adopting common European data protection standards, the EU was able to harmonize conflicting national laws that were emerging as a trade barrier, inhibiting commerce in Europe. For this reason, GDPR and its predecessor were viewed as internal market instruments, facilitating the creation of a digital, single market by allowing an unhindered flow of data within the entire common market. Due to its ability to bridge the information asymmetry between supply and demand, data now has an economic value. When platforms compile personal data, they gather preferences and interests, which allow companies to exert a targeted action on the consumer through advertising. Algorithms classify, reference, and prioritize the preferences of individuals to better predict their behavior. Via free access to platforms in exchange for the collection of personal data, they make the content non-rival. Thus, the intangibility of content tends to give a collective natural aspect to this information accessible to everyone, to benefit public good by creating a digital public space. The McKinsey Global Institute Report (2014) notes five broad ways in which leveraging big data can create value for businesses: Creating transparency by making data more easily accessible to stakeholders with the capacity to use the data, Managing performance by enabling experimentation to analyze variability in performance and understand its root causes, Segmenting populations to customize products and services, Improve decision making by replacing or supporting human decision making with automated algorithms, Improve the development of new business models, products, and services. In 2011, the Boston Consulting Group estimated that personal data collected in Europe was worth 315 billion euros. Network effect The Network effect occurs when the value of a product or service to the user increases exponentially with the number of other users using the same product or service. For instance, WhatsApp provides a free communication platform with friends and contacts. The utility to use it relies on the fact that a substantial part of or friends and colleagues are already users. Multi-sided market The Digital market can be labeled a ‘multi-sided’ market. The notion developed by French Nobel prize laureate Jean Tirole is based on the idea that platforms are ‘two-sided’. This can explain why some platforms can offer free content, with customers on one side and the software developers or advertisers on the other. On a market where multiple groups of persons interact through platforms as intermediaries, the decisions of each group affect the outcome of the other group of persons through a positive or negative externality. When the users spend time on a page or click on links, this creates a positive externality for the advertiser displaying a banner there. The digital Multinational enterprises (MNEs) do not collect revenue from the user side but from the advertiser side, thanks to the sale of online advertisement. Tendency to oligopoly and monopoly formation The outcomes of these intertwined and combined effects tend to lead to the formation of dominant market positions, also called digital monopoly or oligopoly. In this sense, digital platforms such as the GAFA (Google, Apple, Facebook, and Amazon) can be considered as first movers - large companies that introduce a service or a product on an immature market, allowing that company to establish strong brand recognition and service loyalty. Response Given its expected broad impact, traditional firms are actively assessing how to respond to the changes brought about by the digital economy. For corporations, the timing of their response is of the essence. Banks are trying to innovate and use digital tools to improve their traditional business. Governments are investing in infrastructure. In 2013, the Australian National Broadband Network, for instance, aimed to provide a 1 GB/second download speed fiber-based broadband to 93% of the population over ten years. Some traditional companies have tried to respond to the regulatory challenge imposed by the Digital economy, including through tax evasion. Due to the immaterial nature of digital activities, these digital multinational enterprises (MNEs) are extremely mobile, which allows them to optimize tax evasion. They can carry out high volumes of sales from a tax jurisdiction. Concretely, governments face MNE fiscal optimization from companies locating their activity in the countries where tax is the lowest. On the other hand, companies can undergo double taxation for the same activity or be confronted with legal and tax vagueness. The Conseil National du Numérique concluded that the shortfall in corporate tax gain for Apple, Google, Amazon, and Facebook was worth approximately 500 million euros in 2012. According to 55% of businesses surveyed in the European Investment Bank's Investment survey in 2021, the COVID-19 pandemic has increased the demand for digitalization. 46% of businesses report that they have grown more digital. 34% of enterprises that do not yet utilise advanced digital technology saw the COVID-19 crisis as a chance to focus on digitisation. Energy use The Digital Economy uses a tenth of the world's electricity. The move to the cloud has also caused the rise in electricity use and carbon emissions. A server room at a data center can use, on average, enough electricity to power 180,000 homes. The Digital Economy can be used for mining Bitcoin which, according to Digiconomist, uses an average of 70.69 TWh of electricity per year. The number of households that can be powered using the amount of power that bitcoin mining uses is around 6.5 million in the US. Cashless society A cashless society describes an economic state in which transactions no longer use physical currency (such as banknotes and coins) as the medium. Transactions which would historically have been undertaken with cash are often now undertaken electronically. This has become a topic of increasing interest in today's society as digital or virtual currencies for transactions become more common. This is also an important part of the digital economy. One nation moving towards achieving this idea of cashless society is El Salvador, who became the first country to adopt Bitcoin. Bitcoin is a cryptocurrency that only exists digitally, as legal tender in the country. While technical issues occurred in the rollout, El Salvador's President Nayid Bukele, a supporter, noted that it would generate investment interest in the country, as well as providing access to approximately 70% of citizens who lack access to "traditional financial services." The effect of increased energy use from mining bitcoin on carbon emissions in El Salvador has yet to be seen. EU Digital area Remaining barriers to fulfill the Digital Single Market The Digital Single Market (DSM) was included as part of the Single Market Act initiatives adopted by the European Commission (EC). The question had already come up earlier in 1990 and was brought up again later in 2010, emerging at a sensitive moment in the post-crisis of 2008, and used as a catalyst for action. The crisis created opportunities to place the Single Market upfront in the European agenda and was aimed to resolve two issues: financial supervision and economic coordination. This gave a new dimension to the Market. The proposal for the DSM had been made under the strategy of the Commission entitled "Digital Agenda for Europe" in the political guidelines of the second Barroso Commission and pointed out the need to eliminate barriers in order to implement the European Digital Market as an attempt to relaunch the Single Market. This strategy was similar to the one used for the Internal Market in 1985 and focused on one of the weaknesses of the latter, namely the fragmentation of the national digital market. Building on the Monti report, the communication 'Towards a Single Market Act' detailed 50 proposals to reform the SM by the end of 2012. But the DSM was only adopted in 2015 and the proposal for a directive of the European Parliament and the Council was made in September 2016. The DSM is presented as a key priority in the economy of Union, even if there were several attempts to deepen the integration, there are still obstacles remaining. The creation of the DSM constitutes a catalyst to resolve several issues, and was supposed to have a widespread multiplier effect throughout sectors across the EU. The EU Commission faced several obstacles. The commission acts in a way to deeply transform the Single Market. However, the EC lacked political support to enhance the impact of its decision. The issue of the low salience was a causal factor explaining the limits of the commission's commitment to reform the single market. Even though the member states approved the DSM, and the definition for the DSM was accepted by European institutions as a key priority, only one proposal was adopted at the end of 2012. Despite being a priority in the SMA I & II, legislative initiatives failed due to the high cost of implementation measures. Also, there were its potential ‘blockbuster for economic gains’  and the protest of citizens against sovereign debt countries' rescue and bank bail-outs. The slow adoption of the proposal is partly due to member states’ protectionist temptations after the economic crisis. Each state wanted to put forward its preferences and legislation concerning this field. With regard to artificial intelligence (AI), the Commission adopted various initiatives with no meaningful coordination. The more pervasive the digital ecosystem becomes, the more sector-specific regulatory framework may need to be merged into general regimes. Though the Commission used the crisis as a window of opportunity, it did not allow it to go further in implementing a high transformation of the Single Market. The crisis context pushed the political actors to move forward to better manage the crisis, but did not permit it to fully implement the DSM. Current challenges One of the key priorities of the EU is to guarantee fair competition. Yet, within the Digital Market, the competition may be distorted. With more exertion of network effects comes higher barriers to entry (difficulty for a new entrant to enter the market and compete) in the market. Vertical or horizontal mergers and acquisitions take place in closed ecosystems. In order to limit this problem in the digital ecosystem, the EU aims to qualify certain firms as either as an "abuse of dominant position" or a "cartel" which are against the competition prosperity within the Single Market. Digital companies such as the GAFA prosper thanks to their various free services that they make available to consumers, which appear beneficial for consumers, but less so for firms in potential competition. It my be difficult for regulators to sanction firms such as GAFA, due to the jobs and services they provide worldwide. Challenges for the regulator Certain challenges may exist for regulators. One example is in identifying and defining platforms. Member states lack coordination, and may be independent of the regulator, who can not have a global vision of the market. Also, tax evasion of digital MNEs has become a growing concern for most of the European governments, including the European Commission. Attracting foreign investment is less and less seen as a relevant reason to implement tax cuts. Aside from the fiscal revenue shortfall, this issue has taken a political turn in recent years since some people and politicians feel that, in a time of financial crisis, these highly profitable firms do not contribute to the national effort. Strength within the EU digital policy The Digital Market is characterized by its heterogeneity. The European Market is in a difficult position to compete with other advanced countries within the Digital World (such as US or China). There are currently no European digital champions. The European Digital Market is divided in regulations, standards, usages, and languages. The member states cannot meet the demand, or support innovation (R&D), due to the fact that the digital environment is by nature global. As noted by the European parliament, taxation on Digital Market could bring about 415bn euros to the EU economy, and be considered as an incentive to further deepen the EU integration (EP opinion's 2014). Mechanisms of control The EU controls ex-post (in the case of abuse of dominance for example) and seems to be very cautious in term of concurrence (exclusive competence). The EU sanctions cartels’ behavior and examines mergers in order to preserve competition and protect small and medium enterprises (SMEs) entering the market. Within the digital market, mergers often create digital firm dominance, thus possibly preventing European equivalents. Moreover, regulation could in theory protect people working in the digital sector or for the digital sector (such as Uber drivers, a case recently in France), which could present opportunity. However, the EU may need to be cautious with regulation in order to create barriers at the market entry. European Commission versus Google In 2017, the EC fined Google €2.42 billion for abusing its dominant position as a search engine by giving an illegal advantage to Google Shopping. The EC aimed to pave the way to relieve firms suffering from its abuse of dominant position. Moreover, it sought to prove that the EC's strategy does works and companies may be fined at high rates. Juncker Commission The Digital Economy has been a concern for the Juncker Commission concern since the 1st Barroso Commission. Yet, it is only under the Juncker Commission that the strategy of the DSM was adopted on 6 May 2015 as it was ranked as the second priority out of the 10 priorities for the new Commission's mandate. Throughout this document, the DSM emphasized 3 policy pillars: improving access to digital goods and services, an environment where digital networks and services can prosper, digital as a driver of growth. As a key priority for the newly President-elect Juncker, he made Andrus Ansip, the vice-president of the Commission, in charge of the DSM. The decision to approach the DSM from a different point of view is also because the digital space is in constant evolution with the growing importance of online platform and the change of market share. The DSM was a priority because of its economic importance; the total of EU e-commerce reached 240 billion € in 2011, and out of that 44 billion were cross-border trade between member states. Within the new commission In 2020, the digital economy continues to be a top priority for the EC, and belongs once again to the agenda of the Commission president. Frans Timmermans has been designated to be the vice president in charge of one of the six priorities of the EC, called "A Europe fit for digital age". The priority is elaborated as follow, EC is working on a digital transformation that will benefit to everyone. These goals are set to open up new opportunities for businesses, to boost the development of trustworthy technology, foster an open and democratic society, enable a vibrant and sustainable economy, and help fight climate change and achieve the green transition. The strategy of digital economy is included in a wider strategy for the future of Europe. However, as explained on the EC's website, the aim to become a global role model for the digital economy fit within the EU's goals for decades, as it is the aim in the environmental field. However, the EU had to review its aim in this field, and becomes a 'Leadiator'. It is possible that in the Digital Economy, the EU has to behave and evolve in the same way, because the champions of the digital sector aren't European, which creates a handicap in the way the EU refrain from legislating. The EU cannot restrict the offer to its citizens, because digital leaders are not mainly Europeans. One objective of the single market is to make available the better quality at the better price, and propose a better choice to its citizens. Conclusion As explained earlier, the digital economy is based on computing technologies. A rise in new businesses results in greater business connectivity throughout the world. What's referred to as a 'new type of economy' emerges. The rapid spread of ICT all around the world has led to the development of new kind of products and services, that has changed the way business is done. The Digital Economy represents today 15% of the global world GDP. It is relying on personal data, which has been regulated by the EU's directive of 1995, which had the goal to integrate EU within the Digital market. Digital Single Market has been for long a priority for the EU and has beneficiated of the 2007 crisis as a window of opportunity to act. However, it's said the mandate of the EC is thin, to the heterogeneity of the market, and the fact that the EU has to act ex-post. As a result, member states lack coordination. The goals of the Single Market concerning consumers is to offer a panel of choices at a better price. Yet, the leaders of the Digital market may not necessarily be European. Due to the network effect, barriers to European businesses that want to enter within the market remain even higher. Dominant position harbored by US big tech platforms do not give a wide possibility of maneuver combined with the volatility of the market. Critics/debate Rise of intangible capitalism The digital economy is also qualified as "intangible capitalism" which fosters inequality and social division. In 2017, Haskel and Westlake published "capitalism without capital" which raises concerns about policymakers’ inability to tailor from the transition of the traditional economy to the New Economy based on intangible assets. From the mid-2000s onwards, companies have been investing more in ‘intangibles’ such as branding, design, and technology than they have in machinery, hardware, or property. Many businesses' key assets are primarily software and data (such as Uber), rather than physical. Other business rely on branding to help them stand out from the crowd. Pharmaceutical companies have vast budgets for marketing as well as research and development. In traditional production, marginal cost decreases with volume due to economies of scale and learning curve effects. For digital products and services, such as data, insurance, e-books, and movies, this effect is magnified, because after the first unit, production costs for each additional unit are virtually zero. As the proportion of the world's economy that does not fit the old model keeps getting larger, it has implications for a wide range of policies. The intangibility of assets may widen the gap between small and medium enterprises (SMEs) and multinationals enterprises (MNEs). On the one hand, the current bank system struggles to value and monitor immaterial assets. Historically, when a company went bankrupt, banks could recover their money by selling the physical assets such as buildings, machinery, etc. Yet, if the intangible assets drop, those assets can not be sold easily as the value of the company goes down. As a result, SMEs are more reliant on venture capital which is different from bank financing. The easier access to resources allow MNEs to benefit from synergies of the intangible assets. For instance, in creating the iPod, Apple combined MP3 technology with licensing agreements, record labels, and design expertise to produce the product. This ability to combine technologies and then scale up to help these companies to increase their dominant position on the market. Exploitation of labor forces Expansion of Global value chains The digital economy has accelerated the spread of global value chains in which MNEs integrate their worldwide operations. These advances, coupled with liberalization of trade policy and reduction in transportation costs, have expanded some advantages of businesses in all sectors. For example global value chains in which production processes can be geographically dispersed in locations around the world to take advantage of the features of local markets. It is easier for firms to implement their activities where there are low wages and to coordinate their activities from countries with high wages. Bypassing labor laws The rise of online platforms raises concerns in terms of legal questions about social security and labor law. Since the 2007-2008 financial crisis, there is an increase in 'uberization' of work. As within the company that gives its name to this phenomenon (Uber), workers are defined as ‘independent workers’ (with temporary, off-site, autonomous contracts) which challenges the application of labor and occupational health and safety law. As a result, online platforms encourage the flexibilization of jobs and a higher volatility of the labor market, as opposed to traditional companies. Gig economy companies such as Deliveroo and Uber hire self-employed drivers who sign a contract with the digital platform while the way they work is similar to a regular employee statute. Yet, for the first time, in March 2020, France's top court (Cour de Cassation) ruling acknowledged that an Uber driver could not qualify as a ‘self-employed’ contractor because he could not build his clientele or set his prices, establishing a relation of a subordinate of the company. Intensification of the global competition for human resources Digital platforms rely on 'deep learning' to scale up their algorithm's capacity. The human-powered content labeling industry is constantly growing as companies seek to harness data for AI training. These practices have raised concerns about the low-income and health-related issues of these independent workers. For instance, digital companies such as Facebook or YouTube use ‘content monitor’, contractors who work as outside monitors hired by a professional services company subcontractor, to monitor social media to remove any inappropriate content. Thus, the job consists of watching and listening to potentiallydisturbing posts that can be violent or sexual. In January 2020, through its subcontractor services society, Facebook and YouTube have asked the ‘content moderators’ to sign a PTSD (Posttraumatic Stress Disorder) disclosure after alleged cases of mental disorders witnessed on workers. See also Knowledge economy References Further reading Petrenko Sergei. Cyber Security Innovation for the Digital Economy, Cyber Security Innovation for the Digital Economy: A Case Study of the Russian Federation, (Hardback) and (Ebook) © 2018 River Publishers, River Publishers Series in Security and Digital Forensics, 1st ed. 2018, 490 p. 198 illus. Huws, Ursula. iCapitalism and the Cybertariat - Contradictions of the Digital Economy, in Monthly Review, Volume 66, Issue 08 (January 2015) Cashless society Economic systems Information economy
62796857
https://en.wikipedia.org/wiki/Anjaam%20Pathiraa
Anjaam Pathiraa
Anjaam Pathiraa () is a 2020 Indian Malayalam-language crime thriller film written and directed by Midhun Manuel Thomas. The film stars Kunchacko Boban, Sharaf U Dheen, Sreenath Bhasi, Jinu Joseph and Unnimaya Prasad. Sushin Shyam composed the music, while Shyju Khalid was the cinematographer. The plot follows a squad of police sleuths assisted by a consultant criminologist, Anwar Hussain, as they attempt to unveil the identity of a faceless killer who employs brutal tactics to slay his victims. The film was dubbed in Telugu as Midnight Murders and released on the OTT platform Aha. The film was remade in Bengali in 2021 as Mukhosh. The film went on to emerge as the highest-grossing Malayalam film of the year. Following its success, the director has announced a sequel named Aaram Pathiraa (). Plot Anwar Hussain is a psychologist who works as a consulting criminologist in Kochi City Police on the special recommendation of his friend ACP Anil Madhavan. One night, DYSP Abraham Koshi is abducted and his corpse is left in a gypsum field with eyes and heart eviscerated. The investigation is headed by DCP Catherine. The autopsy report shows zolpidem presence in the body, but no signs of coercion. In a few days, a second police officer goes missing, his body is found in the backyard of the police station, along with a Lady Justice figurine. Anwar finds a similar artefact in the previous crime scene from the forensic photos. Autopsy results are the same as previous. The station's CCTV footage has tampered, indicating the killer is also a security hacker. Anwar employs hacker Andrew for cyber investigation. Catherine implements a plan to lure the killer in which two officers should pair and patrol the nights in 100-meter distance. During duty, Catherine's driver Paulson is kidnapped. His scooter's camera footage shows a wolf-masked man and another man's voice, indicating they work as a team. Paulson's body is delivered to Anwar's house in a refrigerator box. The fingerprint of a cocaine dealer, Shameer, is found. They track him to a cottage where they find his mummified corpse. It was a decoy. Anwar accidentally discovers a signature imprinted on the figurine, which turns out to be that of its sculptor, Sudhakar Devalokam. On questioning, Sudhakar explains that he made five such figurines for two men, who blinded him after he delivered it. From that, Anwar deduces there will be two more murders. On another day, CI Sharathchandran is abducted after a struggle at his home, where in addition to the figurine, they also find a fidget spinner. His body is dropped in a Kochi Metro train by a woman. With that, Catherine is replaced with ACP Prakash Seetharam. Sharathchandran had regularly taken a prescription drug named Tenormin. Anwar theorises that the killers inject hypnotic drug Zolpidem in victim's body somehow before kidnapping and hypnotize them for taking away; Tenormin is a counter-drug against Zolpidem, so Sharathchandran's hypnosis was unsuccessful, hence the struggle. Anwar recalls seeing a photo of a similar spinner in the police file for a homicidal psychopath and hacker named Simon. While Simon is recorded dead during an explosion in prison, Anwar is not convinced. Andrew recognises Simon as the one who cross-dressed as the woman on the train. Anwar reports his findings but is informed that Simon and his crime partner Aravindan's bodies were found that morning with a suicide note confessing to the murders left along with the last figurine. Anwar is unconvinced and believes it is a decoy to stop the investigation. Anwar is called by Sudhakar after a man comes to his workshop for a sixth figurine. He has made a facial composite. Anwar identifies him as psychologist Dr Benjamin Louis since he always recommends his psychology books for referring when he goes to be a guest lecturer to psychology colleges. From Benjamin's foster parent Sudhevan, Anwar learns that when Benjamin was a teenager, his sister Rebecca was impregnated by a priest named Bennet Franko. When their father Louis reported the crime, Anil and Sharathchandran framed him for raping his daughter after taking a bribe from Father Bennet. Rebecca was placed in a distant convent, and their father committed suicide in prison. Later, Benjamin went to the United States for higher studies and came back. Anwar deduces that the first three murders were random and the actual targets are Sharathchandran and Anil. He goes to warn Anil but finds only his deserted car. From a nearby quadcopter, Anwar finds footage of Benjamin's van, Andrew traces it to a pig farm. Anwar and two officers intercepts Benjamin before he can kill Anil. Benjamin knocks them down and chokes Anwar, but is saved by Catherine and Benjamin is arrested. On questioning, Benjamin does not disclose his motive. Anwar deduces that the burned corpse misidentified in prison as Simon's was actually Bennet's. While in transit for presenting him in court, his escort officers kill Benjamin in a fake encounter. At the same time, a woman visits Anil at his apartment. Anwar enters when she leaves and finds Anil dead. Through the window, Anwar calls out the name Rebecca,to which she turns back and looks at him and credits began to roll. Cast Kunchacko Boban as Dr. Anwar Hussain Sharaf U Dheen as Dr. Benjamin Louis Sreenath Bhasi as Andrew, a hacker Jinu Joseph as ACP Anil Madhavan Unnimaya Prasad as DCP Catherine Maria Abhiram Radhakrishnan as SI Pradeep Raman Harikrishnan as SI Arun Mathew Divya Gopinath as SI Preethi Pothuval Remya Nambeesan as Fathima Anwar, Anwar's wife Nikhila Vimal as Rebecca Louis (cameo appearance) Jaffar Idukki as Louis, Benjamin's and Rebecca's father Indrans as Ripper Ravi, a serial killer on death row Assim Jamal as Commissioner Hashim IPS Shaju Sreedhar as Paulson, Police constable Sadiq as Dr. Sreekanth, Surgeon Sudheesh as Sudevan Priyanandanan as Sudhakar Devalokam Nazreen Nazar as Mythili Sudhakar Boban Samuel as DySP Abraham Koshy Jaise Jose as CI Sarathchandran Mathew Thomas as Young Benjamin Louis Nandhana Varma as Young Rebecca Louis Sudheer Sufi as Simon Manjooran / Psycho Simon Arjun Nandhakumar as ACP Prakash Seetharam IPS Arun as Fr. Bennet Franco Majeed as Minister Zhinz Shan as Jail police officer Paul D Joseph as Police Constable Preman as Police Constable Dileesh Nair as Cocaine Shameer Siyona as Samskrithi Amina Nijam as Vicky Maria Gilu Joseph as Doctor Rajan Pootharakkal as Aravindan Kartha Santhosh Laxman as Bengali worker Initially, Actor-writer Vineeth Vasudevan (who co-wrote Allu Ramendran) who is also a chakyar koothu artiste was roped in to play Psycho Simon, but later got replaced.He later plays a small role in the movie. Production On 24 June 2019, it was reported that Kunchacko Boban would star in Midhun Manuel Thomas' next directorial and filming was said to begin by late July. Sharaf U Dheen, Unnimaya Prasad, Sreenath Bhasi, Jinu Joseph, Jaffar Idukki and Assim Jamal were also confirmed in the cast, and Sushin Shyam and Shyju Khalid will be the composer and cinematographer, respectively. Midhun said that the film is entirely different from his previous films. The film was formally launched with a pooja ceremony on 17 July 2019. Principal photography began on 1 August 2019, also revealing the title. Music The film score is composed, programmed, and arranged by Sushin Shyam. Release Theatrical The film was released on 10 January 2020. The film was passed with 'U/A' certificate from Central Board of Film Certification. Home media Premiered on 10 April 2020 at 6:30PM in Surya TV. The movie later released on Sun NXT. Reception Critical response Sajin Shrijith of The New Indian Express rated 4 out of 5 and called it "an efficient, well-crafted mystery" and said that it is "director Midhun Manuel Thomas' brillant work so far from his work which is highly focus on content and making." Litty Simon of Malayala Manorama rated 3.5 out of 5 stars calling it "a crafty thriller from Kunchacko & Co." and "Anjaam Pathiraa will definitely make you restless. It is an excellent theatrical experience.In the film which villain is so complicated and line is multilayered from usual genre which thrills you always." Deepa Soman of The Times of India awarded 4 out of 5 and wrote that it was "a decently engaging crime thriller." and appreciated Kunchacko Boban's performance stating that he "gives the right demeanour to Anwar Hussain.Every member of the crew make their better presence in their work which leads to high quality work." Writing for The News Minute, Cris called it "a neatly-packed thriller" and added that "the film moves fast, not wasting time on a song or even a line too many, with noticeable performances and negligible flaws." Baradwaj Rangan of Film Companion South wrote "In the end, Anjaam Pathiraa turns out to be one of those films that takes a hot-button issue and expects us to care simply because it takes that issue up. The drama is undernourished, the thrills are feeble". Box office The film was both commercial and critical success. In the overseas opening weekend, the film grossed US$713,243 from 78 screens in the United Arab Emirates (the best opening of that weekend), US$48,487 (₹34.48 lakh) from 30 screens in the United States, US$6,081 (₹4.33 lakh) from 2 screens in Canada, and A$28,605 (₹13.89 lakh) from 3 screens in Australia. It earned US$19,857 (₹14.2 lakh) in Canada in two weeks, £28,714 (₹26.78 lakh) in the United Kingdom in three weeks, A$85,285 (₹40.98 lakh) in Australia in four weeks, and US$116,759 (₹83.36 lakh) in the US in five weeks. It is the second highest grossing malayalam film of 2020 which collected over ₹50 crore from Worldwide box office behind Shylock. Sequel A sequel titled Aaram Pathiraa was announced by Midhun Manuel Thomas on 10 January 2021. References External links 2020 films Indian films 2020s Malayalam-language films Indian crime thriller films 2020 crime thriller films
3383644
https://en.wikipedia.org/wiki/Ricky%20Ervins
Ricky Ervins
Richard Ervins (born December 7, 1968) is a former American football running back in the National Football League (NFL) for the Washington Redskins and San Francisco 49ers. He played college football at the University of Southern California. High school career Ervins attended John Muir High School in Pasadena, California, where he participated in football, baseball and competed in track and field. College career In college, Ervins' touchdown run won the 1990 Rose Bowl for the USC Trojans and he was the game MVP. He was teammates on a powerful 1990 USC team with Todd Marinovich and Junior Seau. His eight consecutive 100-yard performances were a USC record and this record has since been broken by former Trojans running back Reggie Bush. Professional career Ervins later played for the Washington Redskins as a rookie running back, he was a second on the team in rushing yards with 680 on the season (behind Earnest Byner) and helped the team win Super Bowl XXVI. In the Super Bowl, he was the game's leading rusher, with 72 yards on 13 carries as the Redskins beat the Buffalo Bills 37-24. Ervins received several honors during his rookie year with Washington, including the PFWA all-Rookie, Football Digest All-Rookie, Pro Football Weekly All-Rookie, Football News All-Rookie, College & Pro Football Newsweekly All-Rookie, Quarterback Club’s Rookie of the year, and Washington Redskin Rookie of the year. He stayed with the Redskins until 1994 when he moved to the San Francisco 49ers. He retired in 1995. Personal Ricky was adopted by Tony and Sharon Crutchfield when he was 14 years old. He now runs a business, Xtreme Xplosion, that trains high school athletes in Northern Virginia. On August 22, 2016, The Tournament of Roses announced Bobby Bell, Ricky Ervins, Tommy Prothro, and Art Spander would be inducted into the Rose Bowl Hall of Fame in the Class of 2016. The Rose Bowl Hall of Fame Induction Ceremony then took place on January 1, 2017, outside the Rose Bowl Stadium, one day before the kickoff of the 103rd Rose Bowl Game on Monday January 2, 2017. References External links Ricky Ervins Xtreme xplosion 1968 births Living people American football running backs San Francisco 49ers players USC Trojans football players Washington Redskins players Players of American football from Pasadena, California
2379782
https://en.wikipedia.org/wiki/Integrated%20logistics%20support
Integrated logistics support
Integrated logistic support (ILS) is a technology in the system engineering to lower a product life cycle cost and decrease demand for logistics by the maintenance system optimization to ease the product support. Although originally developed for military purposes, it is also widely used in commercial customer service organisations. ILS defined In general, ILS plans and directs the identification and development of logistics support and system requirements for military systems, with the goal of creating systems that last longer and require less support, thereby reducing costs and increasing return on investments. ILS therefore addresses these aspects of supportability not only during acquisition, but also throughout the operational life cycle of the system. The impact of ILS is often measured in terms of metrics such as reliability, availability, maintainability and testability (RAMT), and sometimes System Safety (RAMS). ILS is the integrated planning and action of a number of disciplines in concert with one another to assure system availability. The planning of each element of ILS is ideally developed in coordination with the system engineering effort and with each other. Tradeoffs may be required between elements in order to acquire a system that is: affordable (lowest life cycle cost), operable, supportable, sustainable, transportable, and environmentally sound. In some cases, a deliberate process of Logistics Support Analysis will be used to identify tasks within each logistics support element. The most widely accepted list of ILS activities include: Reliability engineering, maintainability engineering and maintenance (preventive, predictive and corrective) planning Supply (spare part) support acquire resources Support and test equipment/equipment support Manpower and personnel Training and training support Technical data/publications Computer resources support Facilities Packaging, handling, storage and transportation Design interface Decisions are documented in a life cycle sustainment plan (LCSP), a Supportability Strategy, or (most commonly) an Integrated Logistics Support Plan (ILSP). ILS planning activities coincide with development of the system acquisition strategy, and the program will be tailored accordingly. A properly executed ILS strategy will ensure that the requirements for each of the elements of ILS are properly planned, resourced, and implemented. These actions will enable the system to achieve the operational readiness levels required by the warfighter at the time of fielding and throughout the life cycle. ILS can be also used for civilian projects, as highlighted by the ASD/AIA ILS Guide. It is considered common practice within some industries - primarily Defence - for ILS practitioners to take a leave of absence to undertake an ILS Sabbatical; furthering their knowledge of the logistics engineering disciplines. ILS Sabbaticals are normally taken in developing nations - allowing the practitioner an insight into sustainment practices in an environment of limited materiel resources. Adoption ILS is a technique introduced by the US Army to ensure that the supportability of an equipment item is considered during its design and development. The technique was adopted by the UK MoD in 1993 and made compulsory for the procurement of the majority of MOD equipment. Influence on Design. Integrated Logistic Support will provide important means to identify (as early as possible) reliability issues / problems and can initiate system or part design improvements based on reliability, maintainability, testability or system availability analysis Design of the Support Solution for minimum cost. Ensuring that the Support Solution considers and integrates the elements considered by ILS. This is discussed fully below. Initial Support Package. These tasks include calculation of requirements for spare parts, special tools, and documentation. Quantities required for a specified initial period are calculated, procured, and delivered to support delivery, installation in some of the cases, and operation of the equipment. The ILS management process facilitates specification, design, development, acquisition, test, fielding, and support of systems. Maintenance planning Maintenance planning begins early in the acquisition process with development of the maintenance concept. It is conducted to evolve and establish requirements and tasks to be accomplished for achieving, restoring, and maintaining the operational capability for the life of the system. Maintenance planning also involves Level Of Repair Analysis (LORA) as a function of the system acquisition process. Maintenance planning will: Define the actions and support necessary to ensure that the system attains the specified system readiness objectives with minimum Life Cycle Cost (LCC). Set up specific criteria for repair, including Built-In Test Equipment (BITE) requirements, testability, reliability, and maintainability; support equipment requirements; automatic test equipment; and manpower skills and facility requirements. State specific maintenance tasks, to be performed on the system. Define actions and support required for fielding and marketing the system. Address warranty considerations. The maintenance concept must ensure prudent use of manpower and resources. When formulating the maintenance concept, analysis of the proposed work environment on the health and safety of maintenance personnel must be considered. Conduct a LORA repair analysis to optimize the support system, in terms of LCC, readiness objectives, design for discard, maintenance task distribution, support equipment and ATE, and manpower and personnel requirements. Minimize the use of hazardous materials and the generation of waste. Supply support Supply support encompasses all management actions, procedures, and techniques used to determine requirements to: Acquire support items and spare parts. Catalog the items. Receive the items. Store and warehouse the items. Transfer the items to where they are needed. Issue the items. Dispose of secondary items. Provide for initial support of the system. Acquire, distribute, and replenish inventory Support and test equipment Support and test equipment includes all equipment, mobile and fixed, that is required to perform the support functions, except that equipment which is an integral part of the system. Support equipment categories include: Handling and Maintenance Equipment. Tools (hand tools as well as power tools). Metrology and measurement devices. Calibration equipment. Test equipment. Automatic test equipment. Support equipment for on- and off-equipment maintenance. Special inspection equipment and depot maintenance plant equipment, which includes all equipment and tools required to assemble, disassemble, test, maintain, and support the production and/or depot repair of end items or components. This also encompasses planning and acquisition of logistic support for this equipment. Manpower and personnel Manpower and personnel involves identification and acquisition of personnel with skills and grades required to operate and maintain a system over its lifetime. Manpower requirements are developed and personnel assignments are made to meet support demands throughout the life cycle of the system. Manpower requirements are based on related ILS elements and other considerations. Human factors engineering (HFE) or behavioral research is frequently applied to ensure a good man-machine interface. Manpower requirements are predicated on accomplishing the logistics support mission in the most efficient and economical way. This element includes requirements during the planning and decision process to optimize numbers, skills, and positions. This area considers: Man-machine and environmental interface Special skills Human factors considerations during the planning and decision process Training and training devices Training and training devices support encompasses the processes, procedures, techniques, training devices, and equipment used to train personnel to operate and support a system. This element defines qualitative and quantitative requirements for the training of operating and support personnel throughout the life cycle of the system. It includes requirements for: Competencies management Factory training Instructor and key personnel training New equipment training team Resident training Sustainment training User training HAZMAT disposal and safe procedures training Embedded training devices, features, and components are designed and built into a specific system to provide training or assistance in the use of the system. (One example of this is the HELP files of many software programs.) The design, development, delivery, installation, and logistic support of required embedded training features, mockups, simulators, and training aids are also included. Technical data Technical Data and Technical Publications consists of scientific or technical information necessary to translate system requirements into discrete engineering and logistic support documentation. Technical data is used in the development of repair manuals, maintenance manuals, user manuals, and other documents that are used to operate or support the system. Technical data includes, but may not be limited to: Technical manuals Technical and supply bulletins Transportability guidance technical manuals Maintenance expenditure limits and calibration procedures Repair parts and tools lists Maintenance allocation charts Corrective maintenance instructions Preventive maintenance and Predictive maintenance instructions Drawings/specifications/technical data packages Software documentation Provisioning documentation Depot maintenance work requirements Identification lists Component lists Product support data Flight safety critical parts list for aircraft Lifting and tie down pamphlet/references Hazardous Material documentation Computer resources support Computer Resources Support includes the facilities, hardware, software, documentation, manpower, and personnel needed to operate and support computer systems and the software within those systems. Computer resources include both stand-alone and embedded systems. This element is usually planned, developed, implemented, and monitored by a Computer Resources Working Group (CRWG) or Computer Resources Integrated Product Team (CR-IPT) that documents the approach and tracks progress via a Computer Resources Life-Cycle Management Plan (CRLCMP). Developers will need to ensure that planning actions and strategies contained in the ILSP and CRLCMP are complementary and that computer resources support for the operational software, and ATE software, support software, is available where and when needed. Packaging, handling, storage, and transportation (PHS&T) This element includes resources and procedures to ensure that all equipment and support items are preserved, packaged, packed, marked, handled, transported, and stored properly for short- and long-term requirements. It includes material-handling equipment and packaging, handling and storage requirements, and pre-positioning of material and parts. It also includes preservation and packaging level requirements and storage requirements (for example, sensitive, proprietary, and controlled items). This element includes planning and programming the details associated with movement of the system in its shipping configuration to the ultimate destination via transportation modes and networks available and authorized for use. It further encompasses establishment of critical engineering design parameters and constraints (e.g., width, length, height, component and system rating, and weight) that must be considered during system development. Customs requirements, air shipping requirements, rail shipping requirements, container considerations, special movement precautions, mobility, and transportation asset impact of the shipping mode or the contract shipper must be carefully assessed. PHS&T planning must consider: System constraints (such as design specifications, item configuration, and safety precautions for hazardous material) Special security requirements Geographic and environmental restrictions Special handling equipment and procedures Impact on spare or repair parts storage requirements Emerging PHS&T technologies, methods, or procedures and resource-intensive PHS&T procedures Environmental impacts and constraints Facilities The Facilities logistics element is composed of a variety of planning activities, all of which are directed toward ensuring that all required permanent or semi-permanent operating and support facilities (for instance, training, field and depot maintenance, storage, operational, and testing) are available concurrently with system fielding. Planning must be comprehensive and include the need for new construction as well as modifications to existing facilities. It also includes studies to define and establish impacts on life cycle cost, funding requirements, facility locations and improvements, space requirements, environmental impacts, duration or frequency of use, safety and health standards requirements, and security restrictions. Also included are any utility requirements, for both fixed and mobile facilities, with emphasis on limiting requirements of scarce or unique resources. Design interface Design interface is the relationship of logistics-related design parameters of the system to its projected or actual support resource requirements. These design parameters are expressed in operational terms rather than as inherent values and specifically relate to system requirements and support costs of the system. Programs such as "design for testability" and "design for discard" must be considered during system design. The basic requirements that need to be considered as part of design interface include: Reliability Maintainability Standardization Interoperability Safety Security Usability Environmental and HAZMAT Privacy, particularly for computer systems Legal See also Reliability, availability and serviceability (computer hardware) References The references below cover many relevant standards and handbooks related to Integrated logistics support. Standards Army Regulation 700-127 Integrated Logistics Support, 27 September 2007 British Defence Standard 00-600 Integrated Logistics Support for MOD Projects Federal Standard 1037C in support of MIL-STD-188 IEEE 1332, IEEE Standard Reliability Program for the Development and Production of Electronic Systems and Equipment, Institute of Electrical and Electronics Engineers. MIL-STD-785, Reliability Program for Systems and Equipment Development and Production, U.S. Department of Defense. MIL-STD 1388-1A Logistic Support Analysis (LSA) MIL-STD 1388-2B Requirements for a Logistic Support Analysis Record MIL-STD-1629A, Procedures for Performing a Failure Mode, Effects and Criticality Analysis (FMECA) MIL-STD-2173, Reliability Centered Maintenance Requirements, U.S. Department of Defense (superseded by NAVAIR 00-25-403) OPNAVINST 4130.2A DEF(AUST)5691 Logistic Support Analysis DEF(AUST)5692 Logistic Support Analysis Record Requirements for the Australian Defence Organisation Specifications - not standards The ASD/AIA Suite of S-Series ILS specifications SX000i - International guide for integrated logistic support (under development) S1000D - International specification for technical publications using a common source database S2000M - International specification for materiel management - Integrated data processing S3000L - International specification for Logistics Support Analysis - LSA S4000P - International specification for developing and continuously improving preventive maintenance S5000F - International specification for operational and maintenance data feedback (under development) S6000T - International specification for training needs analysis - TNA (definition on-going) SX001G - Glossary for the Suite of S-specifications SX002D - Common Data Model AECMA 1000D (Technical Publications) - Refer to S1000D above AECMA 2000M (initial provisioning) - Refer to S2000M above DI-ILSS-80095, Data Item Description: Integrated Logistics Support Plan (ILSP) (17 Dec 1985) Handbooks Integrated Logistics Support Handbook, third edition - James V. Jones MIL-HDBK-217F, Reliability Prediction of Electronic Equipment, U.S. Department of Defense. MIL-HDBK-338B, Electronic Reliability Design Handbook, U.S. Department of Defense. MIL-HDBK-781A, Reliability Test Methods, Plans, and Environments for Engineering Development, Qualification, and Production, U.S. Department of Defense. NASA Probabilistic Risk Assessment Handbook NASA Fault Tree Assessment handbook MIL-HDBK-2155, Failure Reporting, Analysis and Corrective Action Taken, U.S. Department of Defense MIL-HDBK-502A, Product Support Analysis, U.S. Department of Defense Resources Systems Assessments, Integrated Logistics and COOP Support Services, 26 August 2008 AeroSpace and Defence (ASD) Industries Association of Europe Integrated Logistics Support, The Design Engineering Link by Walter Finkelstein, J.A. Richard Guertin, 1989, Article References Military logistics Systems engineering
50207902
https://en.wikipedia.org/wiki/Kl%C3%A1ra%20Pollertov%C3%A1
Klára Pollertová
Klára Pollertová (born September 17, 1971 in Prague) is a Czech actress. Biography She comes from a sporting and artistic family, her brother Lukáš Pollert, is an Olympic champion from the 1992 Summer Olympic Games in Barcelona. She made her film debut as a child in 1979 in the eponymous film adaptation of Čapek's Hordubal. In 1983, then appeared in the television series Visitors. In 1989, she played one of the main protagonists in come back into the grave !. Together with her husband, Ivan Trojan, she played his wife in the film One Hand Can not Clap (2003). After graduating from secondary school she studied acting at the Academy of Performing Arts in Prague, where she earned a MgA. degree, and in the theater began to play in 1994, the Drama Studio in Ústí nad Labem. Since 1995 she has been a member of the Prague Theatre under Palmovkou. Personal life Since 1992, her husband is the Czech actor Ivan Trojan, with whom he has four sons Francis (born 1999), Josef (born 2001), Anthony (born 2009) and Vaclav (born 2012). Filmography 1979 – The Hordubal (Hafia) 1980 – The Golden Hen 1981 – The Owl's Nest (tents) 1981 – Counting Sheep (girl Hana's bed) 1983 – Visitors (Ali Lábusová, classmate Adam) 1983 – Keys to the City (Martina) 1985 – The Little Village (Majka Pávková) 1985 – The teeth and hearts (Susan) 1986 – The operation of my daughter (Milena) 1989 – Go back to the grave! (Student Eva Malkovich) 1989 – The Torture of Imagination 1993 – Mussolini – The Road to Power (Carmen) 1993 – The Sea is silent (girl) 1993 – Giovane Mussolini, Il 1994 – Gracious viewer pardons (Therese – Marie) 1995 – Your World (student film) 1995 – The Tale of God's people and pharmacy (Justynka) 1996 – The Prince boxes (Vítěnka) 1996 – It is not like a sponge sponge (Berta) 1996 – As Mr. Pinajs bought from fat cat (Bride) 1996 – The girl with an uncanny memory (Anabel Frídmanová, rising singing star) 1997 – Vojtík and spirits (Kunhutina niece Veronica) 1997 – Dangerous Liaisons (theater recording) 1997 – Cyprian headless and great-great-grandfather (Kunhutina niece Veronica) 1999 – Victims: Assault (Hrabánková, Igor's wife) 2003 – One Hand Can't Clap (Sandra) 2007 – half-life (Karla) 2008 – Love is 2009 – Taste of summer 2009 – Archive (Křižáková woman in the apartment) 2012 – CSI Angel – Part Doll (Marta Skálová) 2012 – In the shadow of (the woman in the ambulance) 2012 – Smichov crying, Brooklyn sleeping (Jana) 2012 – Do not Stop (mother Miki) 2012 – Four Suns (Jerry's mother) Documentary 2003 – Night with Angel 2011 – 13th Chamber Clara Pollertová-Trojanova TV 2012 – Mirror of Your Life 1983 - Expedition Adam 84 Dubbing 1994 – TV movie The Return of Lew Harper – voicing Melanie Griffith (Schuyler Devereaux) 1998 – The TV movie Only love – [dubbing Hallmark] – Marisa Tomei (Evie) 1999 – TV film Obsession Ayn Rand – [dubbing Hallmark] – Julie Delpy (Barbara Brandenová) Theater Činoherní klub 1991 – Utrpení mladého Medarda (Anička) 1991 – Orestés (Hermiona) Narodni theater (Stavovské divadlo) 1991 – Sbohem, Sókrate (Klára) – vystupovala v alternaci s Barbarou Kodetovou DISK Theater 1992 – Její pastorkyňa (Jenůfa) – divadelní představení na DAMU 1993 – Kurva svatá (Tonka) – divadelní představení na DAMU Činoherní studio Ústí nad Labem 1993 – Kurva svatá (Tonka) – divadelní představení na DAMU 1993 – Její pastorkyňa (Jenůfa) 1993 – Tři mušketýři (D'Artagnanova matka / Ketty) 1994 – Valašská čtverylka (Votruba) 1994 – Procitnutí jara (Vendla) 1994 – Na malém dvorci (Aneta Vaševičová) 1994 – Romeo a Julie (Julie) 1995 – Císařův mim (Pamela) Palmovkou Theater 1996–1997 – Život je sen (Rosaura, dáma) 1996–1999 – Jak se vám líbí (Rosalinda) – vystupovala v alternaci 1997 – Nebezpečné vztahy (Prezidentová de Tourvel) – vystupovala v alternaci s Miroslavou Pleštilovou 1997–1999 – Kouzelník z Lublinu (Magda) 1998–1999 – Pitvora (Beatriz – Juana) 2001 – Ubohý vrah (Druhá herečka) Reznicke Theater 1997 – Snídaně u Tiffanyho (Holly) Svandovo Theater 2002–2003 – Čas katů (Mme Charpennet / Popravovaná žena) 2003 – Clavijo (Marie) 2003–2007 – Poručík z Inishmoru (Maired) 2003–2008 – Znalci (Kiki) 2003–2005 – Ženitba (Duňaša) 2004–2005 – Vášeň jako led aneb Myšlenka pana Doma (Félie, milenka pana Doma) 2004–2006 – Vojcek - Der Romantiker (Markéta, manželka Profesora) 2005–2007 – Klářiny vztahy (Klára) 2005–2010 – Žebrácká opera (Jenny) – vystupovala v alternaci s Janou Strykovou 2006–2008 – Mobile Horror (Terhi) 2006–2008 – Periferie (Paní) 2007 – Bomby, prachy a láska! (Gnazezza Monnezza) 2008 – Její pastorkyňa (Kolušina) 2008 – Dorotka (Adéla) 2008–2009 – Oděsa (Pelegeja, místní kurtizána) – roli později převzala Blanka Popková 2009 – Kdo je tady ředitel? (Mette) – roli později převzaly Šárka Urbanovská a Blanka Popková Viola Theater 2008 – Noc po rozvodu (Ona) Leti Theater 2009 – Večeře Broadway Od 2013 – Večírek – vystupuje v alternaci s Miroslavou Pleštilovou References External links 1971 births Czech television actresses Czech film actresses Czech stage actresses Living people 21st-century Czech actresses Actresses from Prague Academy of Performing Arts in Prague alumni Czech voice actresses 20th-century Czech actresses
3144730
https://en.wikipedia.org/wiki/Motorola%20Mobility
Motorola Mobility
Motorola Mobility LLC, marketed as Motorola, is an American consumer electronics and telecommunications company, and a subsidiary of Chinese multinational technology company Lenovo. Motorola primarily manufactures smartphones and other mobile devices running the Android operating system developed by Google. Motorola Mobility was formed on January 4, 2011, after a split of Motorola into two separate companies, with Motorola Mobility assuming the company's consumer-oriented product lines (including its mobile phone business, as well as its cable modems and pay television set-top boxes), while Motorola Solutions assumed the company's enterprise-oriented product lines. In May 2012 Google acquired Motorola Mobility for US$12.5 billion; the main intent of the purchase was to gain Motorola's patent portfolio, in order to protect other Android vendors from litigation. Under Google, Motorola increased its focus on the entry-level smartphone market, and under the Google ATAP division, began development on Project Ara—a platform for modular smartphones with interchangeable components. Shortly after the purchase, Google sold Motorola Mobility's cable modem and set-top box business to Arris Group. Google's ownership of the company was short-lived. In January 2014, Google announced that it would sell Motorola Mobility to Lenovo for $2.91 billion. The sale, which excluded ATAP and all but 2,000 of Motorola's patents, was completed on October 30, 2014. Lenovo disclosed an intent to use Motorola Mobility as a way to expand into the United States smartphone market. In August 2015, Lenovo's existing smartphone division was subsumed by Motorola Mobility. History On January 4, 2011, Motorola Inc. was split into two publicly traded companies; Motorola Solutions took on the company's enterprise-oriented business units, while the remaining consumer division was taken on by Motorola Mobility. Motorola Mobility originally consisted of the mobile devices business, which produced smartphones, mobile accessories including Bluetooth headphones, and the home business, which produced set-top boxes, end-to-end video solutions, cordless phones, and cable modems. Legally, the split was structured so that Motorola Inc. changed its name to Motorola Solutions and spun off Motorola Mobility as a new publicly traded company. 2012–2014: Under Google ownership On August 15, 2011, American technology company Google announced that it would acquire Motorola Mobility for $12.5 billion, pending regulatory approval. Critics viewed Google as being a white knight, since Motorola had recently had a fifth straight quarter of losses. Google planned to operate Motorola as an independent company. In a post on the company's blog, Google CEO and co-founder Larry Page revealed that Google's acquisition of Motorola Mobility was a strategic move to strengthen Google's patent portfolio. At the time, the company had 17,000 patents, with 7,500 patents pending. The expanded portfolio was to defend the viability of its Android operating system, which had been the subject of numerous patent infringement lawsuits between device vendors and other companies such as Apple, Microsoft and Oracle. On November 17, 2011, Motorola announced that its shareholders voted in favor of the company's acquisition by Google for $12.5 billion. The deal received regulatory approval from the United States Department of Justice and the European Union on February 13, 2012. The deal received subsequent approval from Chinese authorities and was completed on May 22, 2012. Alongside the completion of the acquisition, Motorola Mobility's CEO Sanjay Jha was replaced by Dennis Woodside, a former Senior Vice President at Google. On August 13, 2012, Google announced that it would cut 4,000 employees and close one third of the company's locations, mostly outside the US. On December 19, 2012, it was announced that Arris Group would purchase Motorola Mobility's cable modem and set-top box business for $2.35 billion in a cash-and-stock transaction. In May 2013, Motorola opened a factory in Fort Worth, Texas, with the intent to assemble customized smartphones in the US. At its peak, the factory employed 3,800 workers. On April 9, 2014, following the departure of Woodside, lead product developer Rick Osterloh was named the new president of Motorola. Under Google ownership, Motorola's market share would be boosted by a focus on high-quality entry-level smartphones, aimed primarily at emerging markets; in the first quarter of 2014, Motorola sold 6.5 million phones—led by strong sales of its low-end Moto G, especially in markets such as India, and in the United Kingdom—where the company accounted for 6% of smartphone sales sold in the quarter, up from nearly zero. These goals were compounded further by the May 2014 introduction of the Moto E—a low-end device aimed at first-time smartphone owners in emerging markets. In May 2014, Motorola announced that it would close its Fort Worth factory by the end of the year, citing the high costs of domestic manufacturing in combination with the weak sales of the Moto X (which was customized and assembled at the plant) and the company's increased emphasis on low-end devices and emerging markets. 2014–present: Under Lenovo ownership On January 29, 2014, Google announced it would, pending regulatory approval, sell Motorola Mobility to the Chinese technology company Lenovo for US$2.91 billion in a cash-and-stock deal, seeing the sale of $750 million in Lenovo shares to Google. Google retained the Advanced Technologies & Projects unit (which was integrated into the main Android team), and all but 2000 of the company's patents. Lenovo had prominently disclosed its intent to enter the U.S. smartphone market, and had previously expressed interest in acquiring BlackBerry, but was reportedly blocked by the Canadian government due to national security concerns. The acquisition was completed on October 30, 2014. The company remained headquartered in Chicago, and continued to use the Motorola brand, while Liu Jun () — president of Lenovo's mobile device business, became the company's chairman. On January 26, 2015, owing to its new ownership, Motorola Mobility re-launched its product line in China with the local release of the second generation Moto X, and an upcoming release of the Moto G LTE and Moto X Pro (a re-branded Nexus 6) in time for the Chinese New Year. Lenovo maintained a "hands-off" approach in regards to Motorola's product development. Head designer Jim Wicks explained that "Google had very little influence and Lenovo has been the same." The company continued to engage in practices it adopted under Google, such as the use of nearly "stock" Android, undercutting competitors' pricing while offering superior hardware (as further encouraged by Lenovo), and placing a larger focus on direct-to-consumer selling of unlocked phones in the US market (as opposed to carrier subsidized versions). On July 28, 2015, Motorola unveiled three new devices, and its first under Lenovo ownership—the third-generation Moto G, Moto X Play, and Moto X Style—in three separate events. Integration with Lenovo In August 2015, Lenovo announced that it would merge its existing smartphone division, including design, development, and manufacturing, into the Motorola Mobility unit. The announcement came in addition to a cut of 3,200 jobs across the entire company. As a result of the change, Motorola Mobility will be responsible for the development and production of its own "Moto" product line, as well as Lenovo's own "Vibe" range. In January 2016, Lenovo announced that the "Motorola" name would be further downplayed in public usage in favor of the "Moto" brand. Motorola Mobility later clarified that the "Motorola" brand will continue to be used in product packaging and through its brand licensees. The company said that "the Motorola legacy is near and dear to us as product designers, engineers and Motorola employees, and clearly it's important to many of you who have had long relationships with us. We plan to continue it under our parent company, Lenovo." In response to claims by a Lenovo executive that only high-end devices would be produced under the "Moto" name, with low-end devices being amalgamated into Lenovo's existing "Vibe" brand, Motorola Mobility clarified its plans and explained that it would continue to release low-end products under the Moto brand, including the popular Moto G and Moto E lines. Motorola stated that there would be overlap between the Vibe and Moto lines in some price points and territories, but that both brands would have different "identities" and experiences. Moto devices would be positioned as "innovative" and "trendsetting" products, and Vibe would be a "mass-market challenger brand". In November 2016, it was reported that Lenovo would be branding all its future smartphones under the brand "Moto by Lenovo". In March 2017, it was reported that Lenovo would continue to use the "Motorola" brand and logo, citing its recognition as a heritage cellphone brand. Furthermore, Motorola president Aymar de Lencquesaing stated that Lenovo planned to phase out its self-branded smartphones in favor of Motorola. Under Lenovo, Motorola has faced criticism for having an increasingly poor commitment to maintaining Android software updates for its devices, exemplified by negative responses to a 2019 announcement that Android 9.0 "Pie" updates to the Moto Z2 Force in the United States would only be available to the Verizon Wireless model. Products Razr Motorola Mobility's predecessor Motorola Inc released the Razr V3 in the third quarter of 2004. Because of its striking appearance and thin profile, it was initially marketed as an exclusive fashion phone, but within a year, its price was lowered and it was wildly successful, selling over 50 million units by July 2006. Over the Razr four-year run, Motorola sold more than 130 million units, becoming the bestselling clamshell phone in the world. Motorola released other phones based on the Razr design as part of the 4LTR line. These include the Pebl U6, Slvr L6, Slvr L7 (more expensive variant of Slvr L6), Razr V3c (CDMA), Razr V3i (with upgraded camera and appearance and iTunes syncing for 100 songs), V3x (supports 3G technology and has a 2-megapixel camera), Razr V3xx (supports 3.5G technology) and Razr maxx V6 (supports 3.5G technology and has a 2-megapixel camera) announced in July 2006. The Razr series was marketed until July 2007, when the succeeding Motorola Razr2 series was released. Marketed as a more sleek and more stable design of the Razr, the Razr2 included more features, improved telephone audio quality, and a touch-sensitive external screen. The new models were the V8, the V9, and the V9m. However, Razr2 sales were only half of the original in the same period. Because Motorola relied so long upon the Razr and its derivatives and was slow to develop new products in the growing market for feature-rich touchscreen and 3G phones, the Razr appeal declined while rival offerings like the LG Chocolate, BlackBerry, and iPhone captured, leading Motorola to eventually drop behind Samsung and LG in market share for mobile phones. Motorola's strategy of grabbing market share by selling tens of millions of low-cost Razrs cut into margins and resulted in heavy losses in the cellular division. Motorola capitalized on the Razr too long and it was also slow adopting 3G. While Nokia managed to retain its lead of the worldwide cellular market, Motorola was surpassed first by Samsung and then LG Electronics. By 2007, without new cellphones that carriers wanted to offer, Motorola sold tens of millions of Razrs and their offshoots by slashing prices, causing margins to collapse in the process. CEO of Motorola Ed Zander departed for Dell, while his successor failed to turn around the struggling mobile handset division. Motorola continued to experience severe problems with its cellphone/handset division in the latter-2000s, recording a record $1.2 billion loss in Q4 2007. Its global competitiveness continued to decline: from 18.4% market share in 2007, to 9.7% by 2008. By 2010 Motorola's global market share had dropped to seventh place, leading to speculation of bankruptcy of the company. While Motorola's other businesses were thriving, the poor results from the Mobile Devices Unit as well as the 2008 financial crisis delayed the company plans to spinoff the mobile division. Early Android smartphones Droid In 2008, Sanjay Jha took over as co-chief executive officer of Motorola's mobile device division; under Jha's control, significant changes were made to Motorola's mobile phone business, including most prominently, a shift to the recently introduced Android operating system as its sole smartphone platform, replacing both Symbian and Windows Mobile. In August 2009, Motorola introduced the Cliq, its first Android device, for T-Mobile USA. The device also featured a user interface known as Motoblur, which aimed to aggregate information from various sources, such as e-mail and social networking services, into a consistent interface. A month later, Motorola unveiled the Droid, Verizon Wireless's first Android phone, which was released on November 8, 2009. Backed with a marketing campaign by Verizon, which promoted the device as a direct competitor to the iPhone with the slogan "iDon't," "Droid Does," the Droid was a significant success for Motorola and Verizon. Flurry estimated that at least 250,000 Droid smartphones had been sold in its first week of availability. PC World considered the sales figures to be an indicator of mainstream growth for the Android platform as a whole. The Droid was also named "Gadget of the Year" for 2009 by Time. Other Droid-branded devices would be released by Verizon, although not all of them were manufactured by Motorola. In 2010, Motorola released the Droid X along with other devices such as the Charm, Flipout, and i1. In July 2010, Motorola reported that it had sold 2.7 million smartphones during the second quarter of 2010, an increase of 400,000 units over the first quarter. Jha stated that the company was in "a strong position to continue improving our share in the rapidly growing smartphone market and improve our operating performance." In its third quarter earnings report, Jha reaffirmed that the Droid X was selling "extremely well." Atrix 4G, Droid Bionic, XOOM, and Droid RAZR On January 5, 2011, Motorola Mobility announced that the Atrix 4G and the Droid Bionic were headed to AT&T and Verizon, respectively, with expected release dates in Q1 of 2011. The Atrix was released on February 22 as the world's first phone with both a Dual-Core Processor and 1 GB of RAM. The phone also had optional peripherals such as a Multimedia Dock and a Laptop Dock which launched a Webtop UI. On February 24, two days after the release of Atrix, the company released Motorola Xoom, the world's first Android 3.0 tablet, and followed it up shortly afterwards with an update to make it the world's first Android 3.1 tablet. In the fourth quarter of 2011, Motorola unveiled the Droid RAZR, the world's thinnest 4G LTE smartphone at that time at just 7.1 mm. The Droid Razr featured Kevlar backing, the same used in bulletproof vests, and a Gorilla Glass faceplate. The phone was very successful through Verizon Wireless, and many color variants were released. In addition, a Maxx version of the Droid RAZR with an extended battery was released at CES 2012. The Droid RAZR MAXX won CTIA's "Best Smartphone" award. The company also announced new products by late 2011 and early 2012 such as the Xoom 2 tablets, the motoACTV fitness watch with Android, and the Droid 4 with 4G LTE for Verizon Wireless. Though Jha managed to restore some of the lost luster to Motorola Mobility, it struggled against Samsung and Apple. Even among Android manufacturers, Motorola had dropped behind Samsung, HTC, and LG in market share by the second quarter of 2011. This may have been due to the delay in releasing 4G LTE-capable devices, as well as setting the prices of its new products too high. Jha was replaced by Dennis Woodside as CEO by May 2012, when the Google acquisition was complete. Motorola released the Droid RAZR HD (and Droid RAZR MAXX HD) as its 2012 flagship devices, featuring improvements over 2011's RAZR. A lower end RAZR M was released, along with an Intel powered RAZR i. Through late 2012 until 2013's third quarter, no further devices were released, except for the lower end RAZR D1 and D3 devices for Latin America. Google era Moto X (2013-2015) In an August 2013 interview, Motorola Corporate VP of product management Lior Ron explained that the company will focus on the production of fewer products to focus on quality rather than quantity. Ron stated, "Our mandate from Google, from Larry, is really to innovate and take long-term bets. When you have that sort of mentality, it’s about quality and not quantity". Speaking at the D11 conference in Palos Verdes, California, in May 2013, Motorola CEO Dennis Woodside announced that a new mobile device would be built by his company at a 500,000 square-feet facility near Fort Worth, Texas, formerly used by Nokia. The facility will employ 2,000 people by August 2013 and the new phone, to be named "Moto X", will be available to the public in October 2013. The Moto X featured Google Now software, and an array of sensors and two microprocessors that will mean that users can “interact with [the phone] in very different ways than you can with other devices”. Media reports suggested that the phone will be able to activate functions preemptively based on an "awareness" of what the user is doing at any given moment. On July 3, 2013, Motorola released a full-page color advertisement in many prominent newspapers across the US. It claimed that Motorola's next flagship phone will be "the first smartphone designed, engineered, and assembled in the United States". On the same day that the advertisement was published, ABC News reported that customers will be able to choose the color of the phone, as well as add custom engravings and wallpaper at the time of purchase. In early July 2013, the Wall Street Journal reported that Motorola will spend nearly US$500 million on global advertising and marketing for the device. The amount is equivalent to half of Apple's total advertising budget for 2012. On August 1, 2013, Motorola Mobility unveiled the Moto X smartphone. It was released on August 23, 2013, in the US and Canada. On September 5, 2014, Motorola Mobility released the Moto X (2nd generation) smartphone. This continued the trend of the company letting consumers customize their devices through their Moto Maker website, and added new customization options like additional real wood choices and new leather options. The device also got many increases in specs. With a new 5.2 inch (13 cm) 1080p super AMOLED pentile display, a faster 2.5 GHz Qualcomm Snapdragon 801 processor, and an improved 13-megapixel rear camera capable of recording 4k resolution video with a dual LED flash. The device also came with new software features along with new infrared proximity sensors. The Moto X Play and Moto X Style smartphones were announced in July 2015, and were released in September 2015. Many customers who have ordered customized Moto X Pure Editions via Motorola's website have experienced delays receiving their devices. These delays have been attributed to issues including manufacturing issues, lack of parts needed to complete assembly of custom phones (black fronts, Verizon SIM cards and 64 GB versions), a possible redesign due to initial phones having a defect that causes one of the front facing speakers to rattle at high volume and multiple day delays clearing US Customs at FedEx's Memphis, TN hub due to issues related to the import paperwork. . The Moto X Force was launched on October 27, 2015. in the US, it was branded as the Droid Turbo 2, and was the flagship Motorola device of the year, offering Snapdragon 810 processor and 3 GB of memory. Like other Moto X devices, it was customizable through Moto Maker. This is the first Motorola smartphone that features Motorola's "ShatterShield" technology, which consists of two touchscreen elements, reinforced by an internal aluminum frame to make it resistant to bending and cracking, although this does not protect against scratches or other superficial screen damage. The top layer of the display is designed to be user-replaceable. The screen and case also have a water repellent nanocoating to protect the device from liquids that could damage internal components. This was also one of the last Motorola phones to be released under Google's ownership. Droid Mini, Ultra and Maxx Droid Mini, Droid Ultra and Droid Maxx were announced in a Verizon press conference on July 23, 2013. These phones share similar design with the predecessing Droid Razr HD lines in different screen and battery sizes, while all featuring the same Motorola X8 Mobile Computing System as the first-generation Moto X, with exclusive features like Motorola Active Notifications and 'OK Google' on device neural-based voice recognition system. In September 2015, Droid Maxx 2 were launched as Verizon exclusives in the US market, which share the same overall design as the Moto X Style, with Verizon software on board. Unlike its Moto X counterpart, Droid Maxx 2 does not support Moto Maker for further customization. Instead, Moto Shell with 7 different colors was introduced specifically for the device. Moto G On November 13, 2013, Motorola Mobility unveiled the Moto G (1st generation), a relatively low-cost smartphone. The Moto G has been launched in several markets, including the UK, United States, France, Germany, India and parts of Latin America and Asia. The Moto G is available in the United States, unlocked, for a starting price of US$179. The device is geared toward global markets and some US models support 4G LTE. Unlike the Moto X, the Moto G is not manufactured in the United States. On September 5, 2014, Motorola Mobility released its successor to the 2013 version of the Moto G, called the Moto G (2nd generation). It came with a larger screen, higher resolution camera, along with dual front-facing stereo speakers. On July 28, 2015, Motorola Mobility released the third generation of the Moto G series, called the Moto G (3rd generation), in a worldwide press conference in New Delhi, India. It retained the same screen as before but upgraded the processor and RAM. Furthermore, it has an IPx7 water-resistance certification and comes into two variants - 1GB RAM / 8GB ROM and 2GB RAM / 16GB ROM. The device also has the latest (at the time) Android Lollipop OS v5.1.1. In May 2016, Motorola released three fourth generation Moto G smartphones: Moto G⁴, Moto G⁴ Plus, and Moto G⁴ Play. On February 26, 2017, Motorola Mobility released two fifth generation Moto G smartphones during Mobile World Congress: Moto G5 and Moto G5 Plus. On August 1, 2017, Motorola added two 'special edition' models to the Moto G lineup, the Moto G5S and Moto G5S Plus. In May 2018, Motorola released the sixth generation of the Moto G line up in three variants, the G6, G6 Plus and the G6 Play. The G6 and G6 Plus have two rear cameras, capable of taking 4k video (Plus model only), and a larger screen with an aspect ratio of 18:9. The specification ranges from 3/32GB to 6/128GB. In February 2019, Motorola launched the Moto G7 with 3 extra variants, (The Moto G7 Play, Moto G7 Power and the Moto G7 Plus). It has a 6.2 Inch screen, 0.5 inches bigger than The 5.7 inch screen of its predecessor, the Moto G6. The G7 also has two rear cameras, the same as its predecessor. It's said to have a brighter and clearer display than the G6, but with most of the same features. The G7 also has a heavily upgraded mid-range Snapdragon 632 processor compared with the G6 which has a budget Snapdragon 450 Processor. Additionally, the Moto G7 features a hefty 4GB of RAM packed into the budget-friendly device. This is a 1GB upgrade from the G6 which featured 3GB of RAM . Since its Market entry in 2013, the Moto G Series has continued to provide powerful mid-range smartphones for a budget-friendly price. Moto E The Moto E (1st generation) was announced and launched on May 13, 2014. It was an entry-level device intended to compete against feature phones by providing a durable, low-cost device for first-time smartphone owners or budget-minded consumers, with a particular emphasis on emerging markets. The Moto E shipped with a stock version of Android "4.4 "KitKat"." The Moto E (2nd generation) was announced and launched on March 10, 2015, in India. Released in the wake of its successful first generation, the second generation of the Moto E series still aims to provide a smooth experience to budget-oriented consumers. It increased the screen size to 4.5" but kept the resolution at 540 x 960px. It came in two versions, a 3G-only one powered by a Snapdragon 200 chipset and a 4G LTE version powered by a Snapdragon 410 chipset. As before, it shipped with a stock version of the latest (at the time) Android 5.0 "Lollipop". In 2015 Motorola Mobility marketed the 2nd generation Moto E with the promise of continual updates and support, "And while other smartphones in this category don't always support upgrades, we won't forget about you, and we'll make sure your Moto E stays up to date after you buy it." However, 219 days after launch Motorola announced that the device would not receive an upgrade from Lollipop to 6.0 "Marshmallow". It was later announced that the LTE variant of the device would receive an upgrade to Marshmallow in Canada, Europe, Latin America, and Asia (excluding China). China and the US carrier-branded versions of the device remained on Lollipop, with a minor upgrade to version 5.1. However, the 2nd generation Moto E in the USA did continue to receive support via Android Security Patch updates until at least the October 1, 2016 patch for the LTE variant and the November 1, 2016 patch for the non-LTE variant. Google Nexus 6 / Moto X Pro The Nexus 6 was announced October 15, 2014 by Motorola Mobility in partnership with Google. It was the first 6-inch smartphone in the mainstream market, and came with many high-end specs. It was the successor to the Nexus 5, Google's previous flagship phone from their Google Nexus series of devices. Its design was similar to the Moto X (2nd generation) but with a larger display and dual, front-facing speakers rather than the single front-facing speaker on the Moto X. It was the first phone running vanilla Android Lollipop, receiving software updates directly from Google. It was later updated to Android Marshmallow in 2015 and Android Nougat in 2016, though later versions of Android 7.X took some time to arrive, and it never received Android 7.1.2 update, ending its support with Android 7.1.1 in the end of 2017. On January 26, 2015, Motorola Mobility announced that they would sell the Moto X Pro in China. The Moto X Pro was similar to the Nexus 6 in terms of hardware, but excluded all of Google's services and applications. The phone was released in April 2015 with Android 5.0.2 'Lollipop'. However, it never received any Android version update throughout its lifetime despite the same hardware as the Nexus 6. Droid Turbo / Moto Maxx / Moto Turbo The Droid Turbo (Moto Maxx in South America and Mexico, Moto Turbo in India) features a 3900 mAh battery lasting up to two days. Motorola claims an additional eight hours of use after only fifteen minutes of charging with the included Turbo Charger. The device is finished in ballistic nylon over a Kevlar fiber layer and is protected by a water repellent nano-coating. Droid Turbo uses a quad-core Snapdragon 805 processor clocked at 2.7 GHz, 3 GB RAM, a 21-megapixel camera with 4K video, 5.2-inch screen with resolution of 2560 × 1440 pixels. The Droid Turbo includes 32 or 64 GB of internal storage, while the Moto Maxx is only available in 64 GB. In late 2015, Droid Turbo 2 was introduced as a Verizon exclusive, which was a rebrand of Moto X Force. It was the first Droid device to offer Moto Mods for extensive customization, and ShatterShield display technologies. Lenovo era Moto Z and Moto Mods The Moto Z lineup was introduced in June 2016. The smartphone features Motorola's Moto Mods platform, in which the user can magnetically attach accessories or "Mods" to the back of the phone, including a projector, style shells, a Hasselblad-branded camera lens, and a JBL speaker. There were three versions of the original Moto Z. The global flagship model Moto Z, and Moto Z Droid as a Verizon exclusive, were introduced as the thinnest premium smartphone in the world, according to Motorola, and featured a 13-megapixel camera with 4K video, 5.5-inch screen and 4 GB of RAM. and an underclocked Snapdragon 820 chipset at 1.8 GHz, but was able to unlock with root permission granted. Moto Z Play, on the other hand, featured a less powerful processor and a bigger battery. Moto Z Force Droid, only introduced as a Verizon exclusive, featured the Snapdragon 820 chipset with standard frequency, a display with Motorola ShatterShield technology and a 21 megapixel camera. These phones ship with near stock Android 6.0 'Marshmallow' with usual Moto experiences. They were later updated to Android Nougat and Android Oreo in early 2017 and 2018, a substantial delay compared to older Moto X models. Moto Z2 Play was launched in June 2017 with updated Moto experience, a slightly faster processor than Moto Z Play, and Android Nougat. Moto Z2 Force was launched in July 2017, and was the last Motorola flagship phone with up-to-date processor as for now. It was the first non-Google phone to feature 'A/B partition' and 'seamless update' features of Android Nougat, allowing users to install update in the background and finish with just a restart. It was updated to Android Oreo in the end of 2017, which was fairly speedy compared to other Android OEMs. In China, the phones run ZUI from ZUK, a subsidiary of Lenovo instead of near stock Android, and Moto Z2 Force was rebranded as 'Moto Z 2018'. There were also several Moto Mods released in 2017, including a Motorola TurboPower Moto Mod with fast charging capabilities. The Moto Z3 lineup was released in August 2018, consisting of Moto Z3 Play with Snapdragon 6-series SoC and Moto Z3 with last generation 8-series SoC, both supporting all Moto Mods introduced. The 5G Moto Mod was introduced alongside the Moto Z3 lineup, and was a Verizon exclusive at launch in early 2019. It enables 5G connectivity for Moto Z2 Force and later Moto Z devices when the mod is attached on the phone. The phones launched with Android Oreo and was later updated to Android Pie with a similar software experience as earlier Moto Z devices. The Moto Z4 was launched in May 2019 with a 48 megapixel camera sensor and enhanced night sight features, as well as an intergraded fingerprint sensor. It featured a Snapdragon 675 SoC, 4 GB of memory, and launched with Android Pie. Unlike older Moto Z models, this phone was focused on upper mid-range market, and came with a Moto 360 Camera Mod in the box. Moto M The Moto M was introduced in late 2016. It was a mid-range device launched in markets such as mainland China and Hong Kong, Southeast Asia, and South Asia. It featured an octa-core MediaTek processor and 3 or 4 GB of memory depending on storage, and ran near-stock Android. Despite the Moto branding, the bootloader and software update software came from Lenovo directly instead of Motorola, similar to the Moto E3 Power. Moto C The Moto C was announced in 2017 as a low-end device, slotting in below the Moto E as the cheapest in the Moto range. The base model includes a 5-inch display, MediaTek quad-core processor, a 5-megapixel rear-facing camera, and an LED flash for the front-facing camera. It is accompanied by a higher-end model, the Moto C Plus, which features a more powerful processor, larger battery, 8-megapixel camera, and a 720p display. Moto X4 The Moto X4 was introduced in August 2017. It featured a 5.2-inch 1080p IPS display compared to Moto X 2014, and 3, 4, or 6 GB of RAM depending on versions. There were three iterations: a retail one from Motorola, an Android One version from Google, and an Amazon Prime edition. Not to be confused with older Moto X models, this phone does not carry on the older Moto X line, and it did not offer some older Moto X exclusive features such as the Moto Maker. The phone came with Android Nougat 7.1.1 and Moto features, with support of 'A/B Partitions' and 'seamless update'. It was updated to Android Oreo and Android Pie in late 2017 and 2018. It was succeeded by the Moto One lineup since 2018. Motorola One lineup In 2018, Motorola launched the Motorola One lineup as upper mid-range replacements for the Moto X4. In August 2018, the first phones in the lineup, the Motorola One and Motorola One Power were launched. Both phones featured dual camera setup, displays with 'notches' for sensors and front facing camera, and they were all in the Android One program by Google, with guaranteed three-year security updates. The phones launched with Android Oreo 8.1, and were later updated to Android Pie and Android 10 at the end of 2018 and 2019. Motorola One Vision and Motorola One Action were introduced in May and August 2019. These devices all feature Samsung Exynos processors. Moto One Vision featured dual camera setup, with the main shooter being 48 megapixel. It featured a 'hole punch' display for front camera, and 27W TurboPower fast charging. The Motorola One Action featured a triple camera setup with an ultra-wide lens. Both phones were a part of Android One program by Google. They launched with Android Pie, and were updated to Android 10 in early January 2020. Motorola Razr (2020) The Razr is the first foldable smartphone from Motorola, utilizing a similar design as the original Razr. It was introduced in November 2019 and was expected to launch in 2020. It featured mid-range Snapdragon 7-series SoC and Android Pie, with promised updates to Android 10 in 2020. Motorola Edge & Edge+ The Motorola Edge and Motorola Edge+ were introduced in April 2020. The Edge uses the Snapdragon 765G, while the Edge+ uses the Snapdragon 865; both feature Android 10.0, standard 5G and a curved 90 Hz OLED display. The Edge+ is the first Motorola phone to use a 108 megapixel sensor for the main camera with 6K video recording, and marks a return to flagship devices for Motorola. Smartwatches Motoactv Motoactv is a square smartwatch running Android 2.3 released by Motorola Mobility in 2011. It contained a number of hardware features and software applications tailored to fitness training. Moto 360 Moto 360 is a round smartwatch, powered by Google's Android Wear OS, a version of Google's popular Android mobile platform specifically designed for the wearable market. It integrates Google Now and pairs to an Android 4.3 or above smartphone for notifications and control over various features. The second version of this smartwatch was released in 2015. MINNIDIP x RAZR CH(AIR) In August 2020, the MINNIDIP x RAZR CH(AIR) was announced by Motorola. Corporate image Brand licensing The company has licensed its brand through the years to several companies and a variety of home products and mobile phone accessories have been released. Motorola Mobility created a dedicated "Motorola Home" website for these products, which sells corded and cordless phones, cable modems and routers, baby monitors, home monitoring systems and pet safety systems. In 2015, Motorola Mobility sold its brand rights for accessories to Binatone, which already was the official licensee for certain home products. This deal includes brand rights for all mobile and car accessories under the Motorola brand. In 2016, Zoom Telephonics was granted the worldwide brand rights for home networking products, including cable modems, routers, Wi-Fi range extenders and related networking products. See also iDEN WiDEN Mobile phone industry in the United States List of electronics brands List of Motorola products List of Illinois companies Motorola Solutions References External links 2011 establishments in Illinois American companies established in 2011 Corporate spin-offs Manufacturing companies based in Chicago Electronics companies established in 2011 Electronics companies of the United States American brands Chinese brands American subsidiaries of foreign companies Google acquisitions Mobile phone manufacturers Mobile phone companies of the United States 2012 mergers and acquisitions 2014 mergers and acquisitions Motorola
67646766
https://en.wikipedia.org/wiki/Evercast
Evercast
Evercast is a privately held software as a service company that makes collaborative software primarily for the film, television, and other creative industry sectors. Its platform allows remotely located creative teams to collaborate in real-time on video production tasks, such as reviewing dailies, editing footage, sound mixing, animation, visual effects, and other components simultaneously. Its primary users are directors, editors, VFX artists, animators, and sound teams in the film, television, advertising, and video gaming industries. History The company was founded in 2015 by Alex Cyrell, Brad Thomas, and Blake Brinker, and is based in Scottsdale, Arizona. After using the software, film editor Roger Barton joined the company and became a co-founder and investor. In 2020, Evercast won an Engineering Emmy award. Funding In 2020, an unnamed angel investor provided just over $3 million of funding. References Software companies based in Arizona Collaborative software Film editing Web conferencing Impact of the COVID-19 pandemic on science and technology Software associated with the COVID-19 pandemic
357868
https://en.wikipedia.org/wiki/Panzer%20General
Panzer General
Panzer General is a 1994 computer wargame developed and published by Strategic Simulations Inc. (SSI). It simulates conflict during World War II. The designers of Panzer General were heavily influenced by the Japanese wargame series Daisenryaku. Panzer General was a major commercial hit: 250,000 units were sold at full price, and long tail sales continued in the years ahead. It became and remained SSI's best-selling game across all genres, and was named the best-selling computer wargame of all time in 2007. It is the first in the commercially successful Panzer General series. Gameplay Panzer General is a turn-based game, set on operational level hex maps. One plays lone scenarios from either Axis or Allied side and against a computer or human opponent. In Campaign Mode, the player assumes the role of a German Generalissimus against the Allied computer. Panzer General is an operational-level game, and units approximate battalions, although the unit size and map scale from one scenario to the next are elastic. While the names and information for the units are reasonably accurate, the scenarios only approximate historical situations. Its novel feature was to link individual scenarios into a campaign spanning World War II from 1939 to 1945. Units are able to gain experience and become stronger, where success in one battle would award the player prestige to upgrade units, acquire additional units, and select a better scenario for the next battle. The game requires the player to use combined-arms tactics, where each unit is strong against some unit types but very vulnerable to others. Dug-in enemy positions must be softened by artillery, which is vulnerable and needs protection. Before attacking the infantry and anti-tank, one needs first to destroy the enemy artillery that protects them from behind. If no tanks can get within range, one does this mostly by bombers, but then it is advantageous to destroy the air defense units first. The fighter planes must negotiate dual roles: destroying the enemy air force and protecting their own bombers. The player must carefully observe the road system to speed the advance, or may use Bridge engineers to cross the rivers. The game rewards a Blitzkrieg strategy - penetrating deep into the enemy positions while postponing the destruction of some of the encountered enemy units for later. The performance of units is increased by their experience points, which are acquired through combat. In Campaign mode particularly, one then has to protect the experienced units as the most valuable assets. Campaigns Panzer General has 38 scenarios based on real or fictitious battles from World War II. The player can engage in a single battle or a campaign mode. In Campaign Mode, a series of battles unfold as a campaign heads to victory. There is one long campaign as Germany, with five starting locales: Poland (1939); from Poland, through Norway, to the West with possible amphibious invasion in Britain. North Africa (1941); from North Africa to the Middle East. Barbarossa (1941); from the initial invasion of the Soviet border to Moscow. Italy (1943); from the Allied landing on Sicily to the end of the war. Kharkov (1943); from the German spring offensive to the end of the war. The task in most scenarios is to take all objective cities in a given number of turns; taking them at least five turns earlier is considered a major victory. In scenarios from later stages of war, the Germans try to hold positions against a stronger enemy. A typical task is then "hold at least two of our objective cities for 20 turns; for a major victory, hold at least five". All campaigns branch out and end either by the general being sacked for incompetence or end of the war. In Campaign Mode, a major victory could possibly change known historical events. For example, after a major victory over France, the player invades Britain. Later in the game, after a major victory in Barbarossa, the player can convince the German High Command to attack Moscow immediately (which costs him or her much prestige) rather than diverting to Kiev before Moscow. If the player achieves a major victory both in Britain and in Moscow, he or she is allowed to carry out an invasion of the United States and reach Washington. In any other case, he/she must fight well in many battles to get another chance to attack them. If either Britain or USSR survive this attack, they drive the Germans all the way back to Berlin. The best the player can do is to fight well in each battle to have enough prestige for the next one - and to achieve a major victory in the final defense of Berlin. Development The design of Panzer General was based on the Japanese wargame Daisenryaku. The Strategic Simulations (SSI) team had played a Japanese-language version of the game's Sega Genesis release extensively, and were inspired by its streamlined design. Scenario designer Chuck Kroegel later described Panzer Generals structure as "diametrically opposed to the Gary Grigsby type of game that SSI was used to making". He noted that SSI founder Joel Billings was initially hesitant to adopt the design style. The company's Graeme Bayless later wrote of Daisenryaku: "The genius in this game was the fact that it took the highly complex subject (WWII conflict on land) and boiled it down to the pertinent parts". Daisenryaku would go on to serve as the inspiration for the entire Panzer General series. Reception Sales Panzer General was a commercial success. By November 1995, it had sold over 100,000 copies in the United States and 50,000 in Europe. According to William R. Trotter of PC Gamer US, it was particularly popular in Germany. By August 1996, sales had surpassed 250,000 copies. Describing the situation at the time, Computer Gaming World columnist Terry Coleman wrote: "When you consider that a new Windows 95 and Macintosh version has just been released, it seems fair to say that PG will be incontestably the best-selling historical wargame of all time". The following year, T. Liam McDonald of GameSpot reiterated that the game had sold over 250,000 copies, and noted that 60,000 of these sales came from its PlayStation release. In 2007, Retro Gamer dubbed Panzer General "greatest-selling true wargame of all time", and SSI's most successful title across all genres. The magazine noted that it had accrued 250,000 full-price units sold "and many more in the following years". Computer versions Panzer General won Computer Gaming Worlds 1994 "Wargame of the Year" and PC Gamer USs "Best Wargame" awards, and was a runner-up for the former magazine's overall "Game of the Year" award, which went to X-COM: UFO Defense. The editors of Computer Gaming World called Panzer General "the first wargame since 1987 to garner the #1 spot on the CGW Top 100 poll—a fine endorsement for the most exciting wargame in a long time". In 1996, Next Generation listed it as number 51 on their "Top 100 Games of All Time", contending that though the game's interface is simple, the complexity of the strategy is exceptional. In 1996, Computer Gaming World declared Panzer General the 15th-best computer game ever released, and listed the Game Over scene as #8 on its list of "the 15 best ways to die in computer gaming". The magazine's wargame columnist Terry Coleman named it his pick for the best computer wargame released by late 1996. In 1996, Panzer General won the Origins Award for Best Military or Strategy Computer Game of 1995. Console ports Reviews for the 3DO version were generally positive. The four reviewers of Electronic Gaming Monthly remarked that the game's high level of complexity makes it difficult to get into and unappealing to anyone but enthusiasts of the genre, but that the gameplay design is solid and there is a strong dose of variety to the campaigns. A critic for Next Generation argued that while the music and battle animations quickly wear thin, the game allows them to be turned off and "what it lacks in style, it makes up in substance". He made particular note of the impressive depth of the strategy and the ability to control nearly every land and aircraft used in the World War II European theater. Sir Garnabus of GamePro complimented the accuracy of the 3DO port but otherwise panned the game, contending that war strategy games in general are poor due to their minimalist graphics and focus on thinking rather than acting. In their brief review of the PlayStation version, Next Generation remarked: "One of the best PC wargames of all time and still highly thought of, this version doesn't miss a trick". Legacy In 1996, Robert Mayer of Computer Games Strategy Plus argued: Bruce Geryk of GameSpot argued that Panzer General "almost single-handedly resuscitated computer wargaming as a viable genre". Direct sequels Panzer General was the first game in the Panzer General series, which grew in the years after its release. It was followed by Allied General (Panzer General II in Germany), which allows play from the Allied point of view and features four new campaigns. Other sequels include Fantasy General, Pacific General and Star General. Its mainline sequel, Panzer General II (titled Panzer General 3D in Germany), upgraded the interface to use an overlay of photorealistic terrain, and to display different unit facings, resulting in an improved appearance. In the April 2000 issue of PC Gamer, it was voted the 44th best computer game of all time. The series continued with People's General, an upgraded version of Panzer General II based on a World War III scenario between China and the UN. Panzer General also provided a Play by Email (PBEM) system. Panzer General 3D Assault arrived in 1999, with fully 3D graphics. Panzer General III: Scorched Earth was released in 2000, with better graphics and a redesigned interface. See also Panzer Tactics DS References External links Pepa's Panzer General center (fan site) 1994 video games 3DO Interactive Multiplayer games Computer wargames DOS games Linux games Classic Mac OS games Multiplayer hotseat games NEC PC-9801 games Origins Award winners Ubisoft franchises PlayStation (console) games Strategic Simulations games Top-down video games Turn-based strategy video games Video games about Nazi Germany Video games developed in the United States Windows games World War II video games
1647989
https://en.wikipedia.org/wiki/Brandable%20software
Brandable software
Brandable software is typically software created by one company for the purpose of allowing other companies to obtain resell rights or giveaway rights to the software, change the brand associated with it, and sell it as if it were their own. It may also be referred to as private label software or Rebranded Software. Brandable software is usually presented as an alternative to more expensive software development. In most cases, the company providing the software allows many features of the software to be changed by the customer, such as the software name, graphics, installer and website links. There are various uses of brandable software. In many cases, it is placed on the customer's website in order to generate web traffic and increase name recognition. It can also be used as a source of additional revenue for the customer, or as the sole source of revenue for a new business. Customizable components Brandable Software is normally designed so that certain components are customizable with the purchasing company's information. Among the most common are: Program name - The program is usually given a unique name chosen by the purchasing company Company name - The program is usually designed so that it appears to have been written by the purchasing company Name of the author - name of the person who made the program Product link(s) - Links to one or more sites with product or services offers Website - They purchasing company's website is usually displayed in numerous places within the software to promote brand name awareness and provide easy access to their website and other products Payment URL - The program's demo version will normally give the end user the option of opening a web browser directly to a page on the purchasing company's website that will allow them to immediately purchase the full version. Splash screen - The splash screen is displayed when the program is opened. There may be separate splash screens for the program's installer and the program itself. Graphic design for the splash screen and other graphical elements of the brandable software is usually provided and included in the purchase price, but can also be created by the purchasing company Logo - They purchasing company's logo, or a related image, is usually displayed within the software About box - This window usually displays information about the program, version number, and company End User License Agreement - The End User License Agreement (EULA) can normally be customized Readme - Normally a short introduction to the software with contact information or additional instructions. Help file - The help file is usually created so that no reference is made to the program name or company name. This allows it to be used repeatedly for differently branded versions of the software. It is usually possible for the purchasing company to edit the help file. Start Menu - In most cases, a start menu folder with the purchasing company's name is created on the end user's machine ([Company Name] Software). Then sub-folders are created for each of that company's branded programs. This helps to keep all of purchasing company's software folders organized in one central folder. Installer - The installer is usually also branded with the purchasing company's identifying information, graphics and logos. Demo version and full version - In most cases the branded software is delivered with both Demo and Full versions, so that the demo may be distributed free of charge but in limited capacity, and the Full version is only distributed once it has been purchased. Demo version type - A wide variety of demo types are usually available to choose from. These include demos that will run for a certain number of days, uses, time, or a combination of these things. Demo versions may also have a smaller number of features available. Delivery Once the software has been fully branded with the purchasing company's information it is normally bundled into an installer. Some common types are Windows Installer and executable formats. These can be distributed immediately to end-users without any prerequisite software. Resell rights The purchasing company is given the right to resell their branded version of the software, normally at the price of their choosing. Depending on the agreement, the purchasing company may also be allowed to give their branded version of the software away for free or allow other companies to sell it. In most cases the purchasing company is only given the rights to the compiled version of the software, not the software's source code. Giveaway Rights The person or the company has the right to give their branded version of the software away. But does not have the right to resell the software. References Software Software licenses
46989320
https://en.wikipedia.org/wiki/Ferranti%20MRT
Ferranti MRT
The Ferranti Market Research Terminal (MRT) was, arguably, the world’s first application-specific handheld computer. It was designed specifically for the market research sector as a means to augment the regular clipboard schemes that, at the time, were common-place, in social and market research. Despite having an appearance of a calculator built into a clipboard, the reality was that the unit contained a sophisticated form of programmable data-logger that, in response to an interviewer reading questions to the interviewee, had answers digitally recorded (for later uploading and analysis) via pressing appropriate keys on the unit. The unit contained a bespoke operating system to support field based market research. The Ferranti Market Research Terminal (MRT) is also of historical significance to the computing industry since it marked the last original computer design from Ferranti, a long established business (started 1882) that had risen to fame through a collaboration Manchester University to produce the "Mark 1", the world’s first commercial computer and later with Cambridge University producing the "Atlas" and "Titan" computers which, at their peak, held around 25% of the computing market. History Ferranti Plc produced two versions that were labelled the MRT-100 and MRT-200 and based on an original 1978 prototype called the Questronic which had been designed at Sheffield University as part of a collaboration between the Department of Electronics and Electrical Engineering and the Department of Geography. While it is difficult to imagine the context of this product development in the late seventies, it is useful to remember that the IBM PC (which has brought the Windows desktops and laptops that are now commonplace) was only introduced on 12 August 1981. If one ignores calculators, then handheld computers appeared in July 1980 with, perhaps, the first being the Tandy Pocket Computer (Radio Shack TRS-80 Pocket Computer). Later in the same year, Matsushita (now the Panasonic Corporation) produced a handheld computer marketed under the Panasonic and Quasar brand. However, these devices were, of course, general purpose handheld computers unlike the application-specific MRT introduced by Ferranti. A merger in 1989 with the American "International Signal & Control Group" (ISC), led to Ferranti going into bankruptcy in December 1994. With that came the end of the Ferranti MRT and, effectively, Ferranti with its long lineage of UK innovative computing and electronics research, development and manufacture. Product Description A description of the Ferranti MRT was given by Deborah Martell (Cardiff Business School, University of Wales Institute of Science and Technology) in the 1987 book, "Management Information Systems: The Technology Challenge" (edited by Nigel Piercy) in which she described the product as follows: "Accurate survey research is of utmost importance in marketing evaluations and NIT [new information technology] is improving this accuracy all the time. This has been shown by Rowley et-al in their description of the Questronic project at the University of Sheffield. The first commercialised product to evolve from this is the Ferranti Market Research Terminal (MRT). The MRT is a battery-operated, hand-held, data-capture terminal, which is an electronic replacement for the clipboard and pencil used for gathering information from consumers. This innovation offers real advantages for survey research, due to easier processing, since data is stored electronically, it is already in a form which can be directly entered into a computer. Likewise, the elimination of the data preparation stage decreases the errors created by human processing. More importantly survey results can be more topical since the MRT decreases the time delay in preparation. Other benefits accrue from decreased handling and stationary costs and increased interviewer productivity. Indeed the Questronic MRT procedures, with their instant processing abilities, allow more specific monitoring of TV effects. A particular attraction is the removal of the traditional “bottleneck” at the data preparation stage for input into the computer. Successful companies will, in the future, emphasise even more the matching of a product to the consumer requirements. Consequently, the old Henry Ford adage “you can have any colour you want, as long as its black” will be gone forever." Perhaps the only thing to add to Deborah Martell’s description was that the MRT supported the easy transmission of collected data through telephone lines using a portable modem which, at the time, was cutting edge technology. To put these developments into context, the Internet was only commercialized in 1995 when NSFNET was decommissioned and the last barriers to using it for commercial purposes were removed (so the Ferranti MRT was in use for over 10 years before the commercial Internet). See also Ferranti Ferranti Mark 1 Ferranti Titan Ferranti Atlas History of Computing Market Research Opinion Polls Clipboard Mobile Computing References MRT Computers designed in the United Kingdom History of computing in the United Kingdom History of science and technology in England Market research Mobile computers
10452165
https://en.wikipedia.org/wiki/David%20Haussler
David Haussler
David Haussler (born 1953) is an American bioinformatician known for his work leading the team that assembled the first human genome sequence in the race to complete the Human Genome Project and subsequently for comparative genome analysis that deepens understanding the molecular function and evolution of the genome. Haussler was elected a member of the National Academy of Engineering in 2018 for developments in computational learning theory and bioinformatics, including first assembly of the human genome, its analysis, and data sharing. He is a Howard Hughes Medical Institute (HHMI) investigator, professor of biomolecular engineering and founding scientific director of the UC Santa Cruz Genomics Institute at the University of California, Santa Cruz, director of the California Institute for Quantitative Biosciences (QB3) on the UC Santa Cruz campus, and a consulting professor at the Stanford University School of Medicine and the UC San Francisco Biopharmaceutical Sciences Department. Education Haussler studied art briefly at the Academy of Art in San Francisco in 1971 and then psychotherapy at Immaculate Heart College in Hollywood until 1973, when he transferred to Connecticut College, finishing in 1975 with a major in mathematics and minor in physics. He earned an MS in applied mathematics from California Polytechnic University in San Luis Obispo in 1979. Haussler received his PhD in computer science from the University of Colorado at Boulder in 1982. Career and research During summers while he was in college, Haussler worked for his brother, Mark Haussler, a biochemist at the University of Arizona studying vitamin D metabolism. They were the first to measure the levels of Calcitriol, the hormonal form of vitamin D, in the human bloodstream. Between 1975 and 1979 he traveled and worked a variety of jobs, including a job at a petroleum refinery in Burghausen, Germany, tomato farming on Crete, and farming kiwifruit, almonds, and walnuts in Templeton, CA. While in Templeton he worked on his Master's degree at nearby California Polytechnic University. Haussler was an assistant professor in Mathematics and Computer Science at the University of Denver in Colorado from 1982-1986. From 1986 to the present, he has been at UC Santa Cruz, initially in the Computer Science Department, and in 2004 as an inaugural member of the Biomolecular Engineering Department. While pursuing his doctorate in theoretical computer science at the University of Colorado, Haussler became interested in the mathematical analysis of DNA along with fellow students Gene Myers, Gary Stormo, and Manfred Warmuth. Haussler's current research stems from his early work in machine learning. In 1988 he organized the first Workshop on Computational learning Theory with Leonard Pitt. With Blumer, Ehrenfeucht, and Warmuth he introduced the Vapnik-Chervonenkis framework to computational learning theory, solving some problems posed by Leslie Valiant. In the 1990s he obtained various results in information theory, empirical processes, artificial intelligence, neural networks, statistical decision theory, and pattern recognition. Haussler’s research combines mathematics, computer science, and molecular biology. He develops new statistical and algorithmic methods to explore the molecular function and evolution of the human genome, integrating cross-species comparative and high-throughput genomics data to study gene structure, function, and regulation. He is credited with pioneering the use of Hidden Markov models (HMMs), stochastic context-free grammars, and the discriminative kernel method for analyzing DNA, RNA, and protein sequences. He was the first to apply the latter methods to the genome-wide search for gene expression biomarkers in cancer, now a major effort of his laboratory. As a collaborator on the international Human Genome Project, his team, featuring programming work by graduate student Jim Kent, computationally assembled the first draft of the human genome and posted it on the Internet on July 7, 2000. Following this, his team developed the UCSC Genome Browser, a web-based tool that is used extensively in biomedical research and serves as the platform for several large-scale genomics projects. These include the National Human Genome Research Institute (NHGRI)’s ENCODE project to use omics methods to explore the function of every base in the human genome (for which UCSC serves as the Data Coordination Center), NIH’s Mammalian Gene Collection, NHGRI’s 1000 genomes project to explore human genetic variation, and the National Cancer Institute (NCI) Cancer Genome Atlas project to explore the genomic changes in cancer. His group’s informatics work on cancer genomics, including the UCSC Cancer Genomics Browser, provides a complete analysis pipeline from raw DNA reads through the detection and interpretation of mutations and altered gene expression in tumor samples. His group collaborates with researchers at medical centers nationally, including members of the Stand Up To Cancer “Dream Teams” and the Cancer Genome Atlas, to discover molecular causes of cancer and develop a new personalized, genomics-based approach to cancer treatment. Haussler is one of eight organizing committee members of the Global Alliance for Genomic and Clinical Data Sharing, along with David Altshuler from the Broad Institute of Harvard and MIT; Peter Goodhand and Thomas Hudson from the Ontario Institute for Cancer Research; Brad Margus from the A-T Children's Project; Elizabeth Nabel from Brigham and Women's Hospital; Charles Sawyers from Memorial Sloan-Kettering; and Michael Stratton from Wellcome Trust Sanger Institute. He co-founded the Genome 10K Project to assemble a genomic zoo—a collection of DNA sequences representing the genomes of 10,000 vertebrate species—to capture genetic diversity as a resource for the life sciences and for worldwide conservation efforts. Awards and honors Haussler is a member of the National Academy of Sciences, the National Academy of Engineering, and the American Academy of Arts and Sciences and a Fellow of the Association for the Advancement of Artificial Intelligence (AAAI). His awards include the 2011 Weldon Memorial Prize from University of Oxford, the 2009 American Society of Human Genetics (ASHG) Curt Stern Award in Human Genetics, the 2008 ISCB Senior Scientist Award from the International Society for Computational Biology (who also elected him an ISCB Fellow in 2009), the 2005 Dickson Prize for Science from Carnegie Mellon University, and the 2003 Association for Computing Machinery (ACM)/Association for the Advancement of Artificial Intelligence (AAAI) Allen Newell Award in Artificial Intelligence. With Cyrus Chothia and Michael Waterman, Haussler was awarded the 2015 Dan David Prize for his contributions to the field of bioinformatics. References American bioinformaticians 1953 births Living people Fellows of the American Academy of Arts and Sciences Members of the United States National Academy of Sciences Howard Hughes Medical Investigators Human Genome Project scientists University of California, Santa Cruz faculty Fellows of the Association for the Advancement of Artificial Intelligence Fellows of the International Society for Computational Biology University of Colorado alumni Connecticut College alumni
22073882
https://en.wikipedia.org/wiki/Packet%20segmentation
Packet segmentation
In data communications networks, packet segmentation is the process of dividing a data packet into smaller units for transmission over the network. Packet segmentation happens at layer four of the OSI model; the transport layer. Segmentation may be required when: The data packet is larger than the maximum transmission unit supported by the network The network is unreliable and it is desirable to divide the information into smaller segments to maximize the probability that each one of them can be delivered correctly to the destination Protocols that perform packet segmentation at the source usually include a mechanism at the destination to reverse the process and reassemble the original packet from individual segments. This process may include automatic repeat-request (ARQ) mechanisms to detect missing segments and to request the source to re-transmit specific segments. In a communication system based on a layered OSI model, packet segmentation may be responsible for splitting one MPDU into multiple physical layer service data units so that reliable transmission (and potential re-transmission via ARQ) of each one can be performed individually. The ITU-T G.hn standard, which provides a way to create a high-speed (up to 1 gigabit/s) local area network using existing home wiring (power lines, phone lines and coaxial cables), is an example of a protocol that employs packet segmentation to increase reliability over noisy media. See also Fujitsu Ltd. v. Netgear Inc. Packet aggregation Segmentation and reassembly References Packets (information technology)
2067515
https://en.wikipedia.org/wiki/Bj%C3%B8rge%20Lillelien
Bjørge Lillelien
Bjørge Lillelien (29 March 1927 – 26 October 1987) was a Norwegian sports journalist and commentator for the Norwegian Broadcasting Corporation. He was considered to be a popular and engaging personality who commentated on many sports, particularly winter sports and football. It was not unusual for Norwegians to turn off the sound on the TV and listen to him on the radio instead. In September 1981 he came to international prominence when he said “Maggie Thatcher your boys took a hell of a beating” in commentary following Norway's 2–1 victory against England in a World Cup qualifier. Career Lillelien went to Northwestern University to study journalism in 1948. After going back to Norway for his military service, he got a job in the newspaper Fremtiden, after which he joined the Norwegian broadcasting network. In 1963 he began a long-term working relationship with former Olympic champion skier Håkon Brusveen at a cross country ski race at Holmenkollen. This collaboration would last for the rest of Lillelien's working days. Brusveen continued to work for NRK after Lillelien's death. Although privately Lillelien was quite reserved and reflective, he developed a gregarious on-air persona. He said "We shouldn't necessarily make a boring radio programme even if it is a boring football match. We should make a sports show that excites people". He continued to work in broadcasting until just before his death from cancer in 1987. "Your boys took a hell of a beating" Lillelien commentated on radio on Norway's 2–1 victory against England in a World Cup UEFA qualifier in Oslo on 9 September 1981. At the end of the match, alternating between the English and Norwegian language, he exultantly proclaimed amidst scenes of jubilant Norway supporters and players in the football stadium: "We are best in the world! We have beaten England! England, birthplace of giants", before taunting a roll call of English historical figures: "Lord Nelson, Lord Beaverbrook, Sir Winston Churchill, Sir Anthony Eden, Clement Attlee, Henry Cooper, Lady Diana, vi har slått dem alle sammen, vi har slått dem alle sammen! (we have beaten them all, we have beaten them all!). Maggie Thatcher, can you hear me? Maggie Thatcher ... your boys took a hell of a beating! Your boys took a hell of a beating!" A significant portion of his excitement was due to how English football was held in very high regard in Norway at the time, as it was often easier to watch the Football League First Division on television in Norway than in England during the 1970s, with one match broadcast live each weekend, and with score updates from other simultaneous matches. Although the broadcast was initially meant solely for the Norwegian home audience its content soon made its way across the North Sea and became legendary in the history of sports commentary in England. In 2002 it topped a list of the “10 greatest bits of commentary ever” in The Guardian newspaper. Full Norwegian/English version The full Norwegian/English version reads as follows: "Vi er best i verden! Vi er best i verden! Vi har slått England 2-1 i fotball!! Det er aldeles utrolig! Vi har slått England! England, kjempers fødeland. Lord Nelson, Lord Beaverbrook, Sir Winston Churchill, Sir Anthony Eden, Clement Attlee, Henry Cooper, Lady Diana--vi har slått dem alle sammen. Vi har slått dem alle sammen. "Maggie Thatcher can you hear me? Maggie Thatcher, jeg har et budskap til deg midt under valgkampen. Jeg har et budskap til deg: Vi har slått England ut av Verdensmesterskapet i fotball. Maggie Thatcher, som de sier på ditt sprog i boksebarene rundt Madison Square Garden i New York: Your boys took a hell of a beating! Your boys took a hell of a beating!" Translated version The full version, completely translated to English: "We are the best in the world! We are the best in the world! We have beaten England 2-1 in football!! It is completely unbelievable! We have beaten England! England, birthplace of giants. Lord Nelson, Lord Beaverbrook, Sir Winston Churchill, Sir Anthony Eden, Clement Attlee, Henry Cooper, Lady Diana--we have beaten them all. We have beaten them all. "Maggie Thatcher can you hear me? Maggie Thatcher, I have a message for you in the middle of the election campaign. I have a message for you: We have knocked England out of the football World Cup. Maggie Thatcher, as they say in your language in the boxing bars around Madison Square Garden in New York: Your boys took a hell of a beating! Your boys took a hell of a beating!" Parodies and homages In 2002 Lillelien's words were designated the greatest piece of sports commentary ever by The Observer'''s Sport Monthly magazine. Such is its place in British sporting culture that parodies of the commentary have been written to celebrate domestic sporting victories, such as the following when the England cricket team beat Australia to regain the Ashes in September 2005:"Kylie Minogue! Steve Irwin! Holly Valance! Crocodile Dundee! Natalie Imbruglia! Ian Thorpe! Mrs. Mangel! Can you hear me? Your boys took one hell of a beating!"And again, when England's cricket team finally won a game against New Zealand in the 2006/7 one-day series, the BBC's web coverage came up with:"Dame Kiri te Kanawa, Peter Jackson, Neil Finn, Sir Edmund Hillary, Jonah Lomu - we have beaten them all! Helen Clark, can you hear me? Your boys took one hell of a beating!"Following Scotland's 2–1 victory against Norway in Oslo in September 2005, the Daily Record parodied the quote, using known Norwegians:"King Olaf, Roald Amundsen, Liv Ullmann, Edvard Munch, Vidkun Quisling, Thor Heyerdahl, Henrik Ibsen, Edvard Grieg, Monty Python's Norwegian Blue, Morten Harket, and Anni-Frid from ABBA. You boys took a helluva beating!" The Deputy Editor of the Daily Record at the time, Murray Foote, later issued an apology for the inclusion of Vidkun Quisling in the quotation:"While I naturally assumed Quisling was not one of Norway's favourite sons, I was truly unaware of the deep loathing his name engenders amongst your countrymen. Had I been aware of this, I would not have used his name and I apologise unreservedly for the offence it caused." In advance of the Scotland v Norway game at Hampden on 10 October 2008, a parody was created by Tennent Caledonian Breweries (sponsor of the Scottish National Team) for radio and subsequently animated and posted on YouTube. Following Manchester University winning University Challenge against the University of Cambridge Pembroke College in March 2012, a commentator for the Manchester student newspaper observed "We have beaten Pembroke College, Cambridge! Cambridge, birthplace of giants. Rab Butler, Clive James, Ted Hughes, William Pitt the Younger, Eric Idle, Peter Cook, Bill Oddie! Bill Oddie, can you hear me? Your boys took a hell of a beating!" The speech is also parodied in the 2001 British drama film Mean Machine'', starring Vinnie Jones. In the film, at the conclusion of the match between the guards and the prisoners at the fictional Longmarsh prison, in a 3–2 win for the prisoners, the commentators, Bob Likely (Jason Flemyng) and Bob Carter (Jake Abraham) quoted "Guards of Pentonville, Guards of Wandsworth, Walton nick in Liverpool, Policemen of Britain, Traffic Wardens, Parole officers, Wheel clampers, your boys have taken a helluva beating today, a helluva beating!". See also Huang Jianxiang, Chinese football commentator with a similarly memorable match commentary Guðmundur Benediktsson, Icelandic football commentator, also with a similarly memorable match commentary Notes Citations Bibliography (Limited access for URL-addresses only from Norwegian IP addresses.) (Limited access for URL-addresses only from Norwegian IP addresses.) (Limited access for URL-addresses only from Norwegian IP addresses.) External links Observer - The 10 greatest bits of commentary ever Tribute (in Norwegian) on NRK website Newspaper commentary on the Daily Record article and the inclusion of Quisling (in Norwegian) Article describing official apology from Daily Record (in Norwegian, with full apology in English) 1927 births 1987 deaths People from Røyken Norwegian sports broadcasters Norwegian sports journalists Norwegian radio personalities NRK people Norwegian association football commentators 20th-century Norwegian writers
27408374
https://en.wikipedia.org/wiki/Computers%20in%20the%20classroom
Computers in the classroom
Computers in the classroom include any digital technology used to enhance, supplement, or replace a traditional educational curriculum with computer science education. As computers have become more accessible, inexpensive, and powerful, the demand for this technology has increased, leading to more frequent use of computer resources within classes, and a decrease in the student-to-computer ratio within schools. History Origins College campuses used computer mainframes in education since the initial days of this technology, and throughout the initial development of computers. The earliest large-scale study of educational computer usage conducted for the National Science Foundation by The American Institute for Research concluded that 13% of the nation's public high schools used computers for instruction, although non-users still outnumbered users at a ratio of 2 to 1. The study also concluded that computers proved to be very popular with students, and that applications run on early models included sports statistic managers, administration tools, and physics simulators. In 1975, Apple Inc. began donating Apple 1 model computers to schools, and mainframes began to lose their former dominance over academic research. Computer usage continued to grow rapidly throughout this era. In 1977, it was estimated that over 90% of students at Dartmouth College had used computers at some point in their college careers. Walter Koetke, the director of a Lexington, Massachusetts school system commented that, "It's still possible for a student to get through here without using the computer, but he would certainly have to try to do it". In 1983, Drexel University became the first campus to require every student to purchase a laptop. Computer-aided instruction gained widespread acceptance in schools by the early 1980s. It was during this period that drilling and practice programs were first developed for exclusive classroom use. Schools became divided over which computer manufacturers they were willing to support, with grade schools generally using Apple computers and high schools preferring DOS based machines. Hardware shortages in schools became a major issue, leaving many teachers unable to provide enough computers for students to use. Despite this, by 1989 computer usage shifted from being a relative rarity in American public schools, to being present in nearly every school district. Modern Era The early 1990s marked the beginning of modern media technology such as CD-ROMs as well as the development of modern presentation software such as Microsoft PowerPoint. Other computer-based technology including the electronic whiteboard and the laptop computer became widely available to students. Internet technologies were also gaining prevalence in schools. In 1996, Bill Clinton made over $2 billion in grants available in the Technology Literacy Challenge Fund, a program which challenged schools to make computers and the Internet available to every student, connected to the outside world, and engaging. This marked a significant increase in the demand for computer technology in many public school systems throughout the globe. Correlating with the development of modern operating systems like Windows 98 and the continuing support of government funding, the prevalence of educational computer usage boomed during this era. Between 1997 and 1999, the ratio of students to multimedia computers decreased from 21 students per machine to less than 10 students per machine. Colleges began creating specialized classrooms designed to provide students with access to the utilization of the most modern technology available. Classrooms such as the "Classroom 2000" built at Georgia Tech in 1999 which featured computers with audio and video equipment designed to capture detailed recordings of lectures as a replacement for traditional note taking began to become more common. By 2000, the student to computer ratio at some schools in the US decreased to only 1 students per school computer. As collaborative classroom environments became mainstream, more schools as a whole began to invest in powerful networks and faster Internet connections. By 2010, many school districts implemented or encouraged "1:1 learning programs" which would ensure that all students in grade school would be provided with a personal laptop. Computers have significantly changed traditional teaching methodology into a more "hands-on" approach, with Forbes predicting that, "Instead of parking themselves in a lecture hall for hours, students will work in collaborative spaces, where future doctors, lawyers, business leaders, engineers, journalists and artists learn to integrate their different approaches to problem solving and innovate together." Experience has shown, however, that excessive use of computers can be detrimental to a student's basic academic skills. The Organisation for Economic Co-operation and Development has compared average student literacy, numeracy and science skills in 31 countries (considering three large cities in China separately). It also compared the levels of student computer use in those same countries. The study's conclusion, after correcting for social backgrounds and student demographics, was that moderate use of classroom computers produces the best educational outcomes. 21st century Demands The shift in worldwide computer usage and the need for computer skills in today's workforce have pushed the United States government to create guidelines for educators to ensure that students are prepared to meet the demands of the 21st century. The Core Curriculum Content Standards for education are aimed at preparing students with such skills demanded of those entering the 21st century learning environment as well as the 21st century and work environment. Changes such as this, along with the changes in the ways that 21st century learners communicate, have impacted the ways that classroom computers are utilized. Currently, teachers are tapping into the enhanced abilities of current classroom computer technology by utilizing various Web 2.0 tools to enhance their instruction. Such tools are also being used to extend classroom communication outside of the campus through online collaborative tools. Centered primarily on collaboration and sharing, Web 2.0 computer applications encourage student self-expression; interaction with peers, and opportunity for authentic learning experiences Through the implementation and integration of Web 2.0 computer technologies into the classroom setting, authentic and meaningful learning experiences are now able to occur in ways that have been previously unimaginable. Currently, the learning that is taking place is not simply about typical concepts or facts as laid out in school curricula. Instead, it is about the process of building connections As a result, the awareness of the importance and the value of communication is becoming instilled into children. Today, with a single laptop, Webcam, projector, and an Internet connection, a teacher can broadcast and begin collaboration with any other classroom. As groups of learners coalesce around shared passions online, they experience something that is difficult to replicate in physical space. Applications Classroom computer access to Web 2.0 applications such as online learning communities and interactive educational tools offer a more dynamic learning experience, with direct benefits to students. Web 2.0 technologies that are being utilized within the classroom have made it possible for essential learning to be introduced to students during their elementary levels of education and to be refined through their middle, high school and collegiate experiences. As classroom computer technology is being used for different types of communication—for presentation, for class interaction, and for collaboration, students are required to be readers and writers, editors and publishers; and must be willing to collaborate and co-create with others, working closely together to learn even more in the process. Web 2.0 Interaction involves not only sharing ideas or information with someone else, but also receiving feedback. The collaboration engages groups of people in not only sending and receiving feedback but working together for creating, building, and editing, These skills are a necessity for students' futures as they grow and enter the workplace. The goal of using such computer applications is to increase peer-to-peer interactions through digital means, in order to replicate real-life 21st century communication skills. One such technology that has gained significant recognition within K-12 education is the Weblog. Weblogs, or blogs, are frequently accessed on classroom computers due to their positive effects upon students. These online journals are primarily used to support communication in the form of presentation, and they provide a useful tool for class interaction. Weblogs allow students to present their own findings and discoveries to an authentic audience. Receiving feedback about course work not just from your teacher but from your peers, or possibly from the outside world, can be very empowering to students. In their eyes, having the ability to publish their writing on a blog suddenly transforms them into authors and publishers. Blogs make students are of careful and conscientious writing. Students are no longer writing for an audience of one. Instead, their words face an audience of their peers as well as countless others whose primary goal is not only to read, but also to provide commentary, feedback and even to critique and criticize their words. By using classroom computers for such means, students become increasingly more cautious and aware of their grammar, spelling and word-choice knowing that they are the authors of a published piece of writing that will reach a wide audience. Similarly, Wikis are commonly accessed on classroom computers due to their positive impacts. The collaborative environment that wikis facilitate can teach students much about how to work with others, how to create community, and how to operate in a world where the creation of knowledge and information is increasingly becoming a group effort. The implementations and uses of wikis range from the development of group-based writing projects, to collaborative note taking, to brainstorming. Teachers can set up wikis for groups of students, giving them the opportunity to all join in on equal footing to give feedback, to make suggestions and changes, and to jot down ideas. With a wiki, everyone is an author of the wiki at the same time. For example, "Teaching with Technology" is an annual survey that asks teachers about technology in the classroom. Survey findings found it was found that 38.37% of teachers said that technology has had an extremely positive impact on education and 36.63% said that technology has mostly had a positive impact on education. Additionally, Wikis afford students the opportunity to express themselves using multiple modalities. Therefore, children who struggle to express themselves through the written word, are now at ease due to ability to insert music, graphics, video, and photos into their writing. With the help of this technology, students who struggle with language are now able to create multimodal compositions, allowing them to communicate meanings that were once inaccessible or not fully expressed through their printed word. These ongoing collaborative efforts also reinforce the notion of careful, contentious writing. Students' words are no longer for an audience of one, but instead for an audience of countless individuals. The awareness of this global reach reminds students to be cautious with grammar, word choice and style as they know that others will be expanding upon their written ideas. Because of their versatility, Podcasts are also commonly accessed on classroom computers. These downloadable, portable files allow listeners to subscribe to digitally recorded audio clips and replay them at their own convenience. Similarly, the use of Vodcasts has become almost as common, as they allow for students to view digitally recorded video on classroom computers. Accessing Podcasts and Vodcasts on classroom computers allows for differentiated instruction within the classroom environment. These technologies provide students with the opportunity to learn at their own pace and the freedom to go back whenever possible and as frequently as necessary in order to check their understanding. Podcasts are offering ESL students, and those with learning disabilities, the chance to review lectures at their own pace for increased comprehension. In fact, research was conducted to see how technology could help the visually impaired in the classroom. It showed that the visually impaired students found technology is better for them, compared to the regular paper. This demonstrated that they can learn for themselves in a classroom. In this case, the iPad2, helped increase the reading speed of those visually impaired compared to the large printed text on paper. Many ELL and ESL students need additional support with defining commonly used language as well as with the pronunciation of new vocabulary. Other students, such as those with special needs have difficulty reading and reviewing complex texts. Accessing Podcasts and Vodcasts on classroom computers can easily provide helpful tools to address these students' needs. Ineffective use Intrinsic barriers such as a belief in more traditional teaching practices and individual attitudes towards computers in education as well as the teachers own comfort with computers and their ability to use them all result in varying effectiveness in the integration of computers in the classroom. Even with the increased access to computers within education systems the actual use of computers in the class room lacks meaningful activities that involve the students. Teachers also appear to abandon the intended educational uses of the computers purely for communication between staff and parents as well as preparation of more traditional learning materials. A survey collected data from 306 in-service teachers from eight secondary schools. One concept that was found in the research was that for the individual category, which included personal skills and feelings, saw that anxiety and computer self-efficiency had an impact on the teachers ability to teach. Through different organizational skill sets, technology could either be stressful and negative or reinforcing and positive for teachers. References Educational technology Computing and society
3374094
https://en.wikipedia.org/wiki/Zango%20%28company%29
Zango (company)
Zango, (also letter case zango), formerly ePIPO, 180solutions and Hotbar, was a software company that provided users access to its partners' videos, games, tools and utilities in exchange for viewing targeted advertising placed on their computers. Zango software is listed as adware by Symantec, and is also labeled as a potentially unwanted program by McAfee. Zango was co-founded by two brothers: Keith Smith, who served as the CEO; and Ken Smith, who served as the CTO. In April 2009 Zango ceased trading after its banks foreclosed. However, , Hotbar, Seekmo, and ZangoCash, formerly owned by Zango, continue to operate as part of Pinball Corporation. Content Zango's consumer website asserted that the company was "committed to creating a content economy built on a foundation of safe and ethical practices by protecting consumer privacy while offering a fulfilling and high-value content experience." It provided targeted advertising in return for partner companies' content such as sports, comedy, dance, erotic videos, online games, and screensavers. Warner Bros. and others had provided content, but WB terminated its business relationship with Zango after concerns were raised that children viewing Warner Bros. content could be exposed to advertisements for pornography. Undesirable software behaviors StopBadware.org lists a number of undesirable behaviors associated with Zango Easy Messenger, including "behaves as spyware", "automatically runs on startup", "displays pop-up advertisements", "installs adware", and "bundled software cannot be closed". The same site states, "We find that Zango Easy Messenger is not badware, although it does engage in behaviors that users should be aware of." Websense has a Zango-related security advisory dated November 2006, stating that "Websense Security Labs has discovered a number of user pages on the MySpace domain which have videos that look like they are from YouTube. The videos have an installer embedded within them for the Zango Cash Toolbar. When users click on the video, they are directed to a copy of the video, which is hosted on a site called 'Yootube.info.' ... the video downloads and attempts to install setup.exe from Zango Cash." A more detailed analysis of this attack, according to one website, is that "Zango continues numerous practices likely to confuse, deceive, or otherwise harm typical users as well as practices specifically contrary to Zango's obligations under its November 2006 settlement with the FTC." These include failure to include on-screen disclosure of material terms, widespread in-toolbar ads without labeling and hyperlinks, ads for "bogus sites that attempt to defraud users", and third party installations without disclosure. Zango software is listed as adware by Symantec. Computer security company McAfee said in 2005 "this program may have legitimate uses", but described it as a "potentially unwanted program" and an "adware downloader." Automated analysis by McAfee SiteAdvisor in May 2008 reports "629 red downloads" and that during testing, McAfee "found downloads on this site that some people consider adware, spyware or other potentially unwanted programs." History 1999–2001: Pay-to-surf model Keith and Ken Smith, the CEO and CTO respectively, founded ePIPO in 1999. It was one of the first "pay-to-surf" companies, following in the footsteps of AllAdvantage. This business model paid users a minimal amount to surf the Internet while running an application that showed banner ads. Users could also make money by referring new users. After enjoying brief success, the pay-to-surf business model declined with the bursting of the dot-com bubble in 2001. The company, which changed its name to 180solutions, adjusted their technologies in several ways: To show pop-up ads rather than banner ads. To not have any visible GUI. To be bundled with other potentially valuable applications. 2002–2005: Distribution via other programs From 2002 to 2005, 180solutions' applications—such as ncase and 180SA (search assistant)—were distributed via various affiliates. While these affiliates were required by the 180solutions contract and by law to obtain the permission of the user prior to software installation, many did not, resulting in millions of illegal non-consensual installs. Many other affiliates notified users only via the end user license agreement, resulting in millions more arguably legal but essentially non-consensual installs. 180solutions' software showed pop-up ads while a user was surfing the Internet. This software was often bundled with freeware which the user intentionally installed; since permission to install the 180solutions adware was typically hidden in an EULA, most users were unaware they were installing adware. In some cases 180solutions' software was installed as a standalone install. Using this method, an ActiveX prompt simply asked the user to install the software so that they could receive "comparison shopping advertisements." 180solutions contended that the value of the bundled software or the advertisements made up for the inconvenience of the pop-up ads. The value of this trade-off was contested by critics of the adware business model. In 2004, Benjamin Edelman, assistant professor at Harvard Business School and spyware researcher, analyzed the network behavior of 180solutions applications and claimed they redirected commissions to themselves that were properly due to affiliates, and additionally caused merchants to pay commissions when affected users clicked on merchant sites directly. During this time, 180solutions' applications were designed to be difficult to uninstall, requiring the user to download an additional uninstall application made by 180solutions or to use an adware removal tool. In 2005 the software uninstall was standardized to use Windows' "Add or Remove Programs" function, making it easy to uninstall. In 2005, 180solutions implemented a number of initiatives to control the distribution of its software and eliminate non-consensual installs. In March, they acquired one of their distribution partners, a Canadian company called CDT (dba LoudCash, giving them direct visibility into and greater control of many of the formerly "third party" distributors. In June, 180solutions claimed to have re-notified users on its 20-million-user customer base, and implemented a program that notifies all users within 72 hours of install and re-notifies all users every 90 days thereafter. By August, they had filed suit against seven individuals alleged to have illegally distributed its software using a botnet. In November, 180solutions announced an ongoing partnership with the FBI in breaking up a botnet ring in the Netherlands. In December, the company ended distribution of the 180SearchAssistant and closed LoudCash, a remnant from the CDT acquisition. 2006–2008: Downsizing Despite the initiatives of 2005, 180solutions admitted that it was possible for malicious parties to hack their install routines and thus cause fraudulent installs. They claimed that the percentage of fraudulent installs had dropped from over 10% to under 1%. Critics considered that the business model was untenable because fraud against 180solutions which harmed unknowing users via non-consensual installs could never be completely removed. In early 2008, security researchers at Fortinet reported, incorrectly, that the rapidly spreading Facebook widget "Secret Crush" was enticing users to download Zango adware by promising to identify a secret admirer. Zango denied any involvement with the widget, and further investigation by an Infoworld senior writer showed that the Fortinet report was incorrect. On June 16, 2008, the company laid off 68 of its approximately 200 employees; Executive Vice President of Corporate Development York Baur, company co-founder Chief Technology Officer Ken Smith, and company co-founder Chief Information Officer Doug Hanhart also left. Zango said it was narrowing its focus to concentrate on its new product Platrium, a "casual gaming experience" that showed targeted ads, shopping comparisons and search suggestions based on keywords from the user's Internet browsing. On December 15, 2008, Zango closed their Tel Aviv office, which had been the Hotbar headquarters before Hotbar and 180solutions merged, thereby laying off another 50 employees. 2009: Closure In a personal bankruptcy filing following a January 2009 $4.6 million judgment in favor of a former employee, Zango's CEO stated that the company was in default to a bank consortium for over $44 million. On April 20, 2009, industry magazine ComputerWorld reported that Zango had become defunct. A spokesman for video search engine company Blinkx said that although Blinkx had purchased some of Zango's technical assets such as servers, Zango was shut down after the consortium foreclosed. 100% of Zango's assets were sold to Blinkx, in what the Zango CEO characterized as a "fire sale". When asked about employees, a Blinkx spokesman said "As Zango was insolvent, we believe all the employees were laid off." It was later announced that ZangoCash was to become Pinball Publisher Network, part of the Pinball corporation, though it would still be operating under its own name with no change to its services. Litigation Federal Trade Commission charges and settlement On January 23, 2006, a public advocacy group filed two official complaints with the Federal Trade Commission. The Center for Democracy and Technology complaints charged 180solutions with engaging in unfair and deceptive business practices, deliberately duping Internet users into downloading intrusive advertising software. The same year, the Federal Trade Commission charged Zango with "Deceptive Failure to Disclose Adware", "Unfair Installation of Adware", and "Unfair Uninstall Practices" in violation of the Federal Trade Commission Act. Since the FTC ruling, security researchers continued to find Zango involved in problematic installs. In November 2006, Zango settled this complaint, via a consent decree with the FTC, without formally admitting guilt. In the words of the Federal Trade Commission press release, "Zango, Inc., formerly known as 180solutions, Inc., one of the world's largest distributors of adware, and two principals have agreed to settle Federal Trade Commission charges that they used unfair and deceptive methods to download adware and obstruct consumers from removing it, in violation of federal law. The settlement bars future downloads of Zango's adware without consumers' consent, requires Zango to provide a way for consumers to remove the adware, and requires them to give up $3 million in ill-gotten gains." These restrictions were to remain in force for twenty years, and the agreement also requires respondents Keith Smith and Daniel Todd to notify the FTC of the discontinuance of their current business or employment, or of their affiliation with any new business or employment, for ten years. In July 2007, Edelman said, "Zango continues numerous practices likely to confuse, deceive, or otherwise harm typical users as well as practices specifically contrary to Zango's obligations under its November 2006 settlement with the FTC." Other litigation In September 2005, attorney Shawn Collins filed a class action lawsuit against Zango on behalf of three plaintiffs, alleging that Zango deceptively installed spyware on more than 20 million personal computers. The company said its software was voluntarily installed by users who downloaded premium content in exchange for their consent to view advertisements relevant to what they searched for online. During pre-trial discovery, the parties agreed to a dismissal with prejudice, meaning that the suit cannot be brought again by the same complainants. Each side agreed to pay its own fees and expenses. In 2005, Zango sued Zone Labs for labeling it as spyware. In May 2007, Zango filed a lawsuit against PC Tools alleging tortious interference with its business and trade libel, because the PC Tools product Spyware Doctor at that time classified Zango software as malicious and removed it without informing users. Zango dropped the suit after the judge ruled that the suit was "unlikely to succeed on the merits of any of its three causes of action" and refused to grant Zango a temporary restraining order. Also in May 2007, Zango filed in the same court a similar lawsuit against Kaspersky Lab, accusing it of tortious interference, trade libel and unjust enrichment for blocking the installation of Zango software. Kaspersky defended itself by invoking the Communications Decency Act (CDA), saying it was immune from civil liability based on the paragraph of the CDA headed "Protection for 'Good Samaritan' blocking and screening of offensive material". The judge agreed, granting Kaspersky's motion for summary judgment. In 2008, Zango appealed against the ruling. As of April 2008, the National Business Coalition for E-Commerce and Privacy was supporting Zango in the appeal. In June 2009 the court ruled that Kaspersky could not be held liable for any actions it took to manufacture and distribute the technical means to restrict Zango software's access to others. Installation and removal Unusual methods of installation and operation have evolved as software such as Zango is targeted to run on a broader mix of applications and platforms. Zango has stated on their website that any Zango "application" can be completely uninstalled by using the Microsoft Windows "Add or Remove Programs" function. However, Zango may appear as a browser plug-in or in some other form which cannot be removed by the officially recommended procedure. Utilities exist for the purpose of detecting, and in some cases removing Zango and similar software. Some software advisers have recommended that people seeking to remove Zango-like software from their computers search for instructions specific to their particular platform, application, and installation. Variants Hotbar Hotbar (also known as HbTools) is a plugin for Internet Explorer, Microsoft Office Outlook and Outlook Express produced by Zango; Hotbar has closed distribution, but is still available to those who still have the installer. From version 10.2 on, it identifies itself as belonging to Pinball Corp. Hotbar adds a toolbar and the option of extra skins to these programs. It also allows the user to add emoticons to emails created in Outlook or Outlook Express and to check weather reports. Its major revenue comes from the use of pop-ups displayed according to a user's behavior and current URL. The application can show over 15 pop-ups a day, depending on how much Internet browsing has occurred. The last reported usable version is 10.0.357.0, dated October 9, 2007. Currently, it often self-updates to a later version that deactivates the toolbar and Outlook components (as it is completely incompatible with all versions of Internet Explorer after IE7, along with all Windows Mail and Windows Live Mail products), rendering it useless. Hotbar Premium Hotbar as supplied by Zango had an option to turn off the advertisements, for a fee. The fee-based Premium access also unlocked certain content beyond the preview stage in Hotbar's utilities. Grandfathered version 4 Premium accounts had unlimited access, as long as the activation link supplied when purchased was saved. Version 4 of Hotbar is no longer usable, as all content is accessible only to version 10. 365-day accounts were sold, valid only for the Hotbar toolset; if Zango was installed alongside Hotbar Premium. In the overlaid installation, Zango's but not Hotbar's, advertising was shown. Hotbar no longer has premium access available for purchase, so all Premium content is not currently usable. Premium assets may be accessed through examination of linking and related code and insertion into use. Zango itself did not offer Premium access. Hotbar's website does not mention Premium access. Controversy Hotbar is an example of adware due to its banner advertising and use of pop-up windows. The user's browsing habits are also sent to the Hotbar servers with a unique user ID which allows a user's browsing habits to be tracked over an extended period. Spyware researcher Benjamin Edelman defined Hotbar's problems: In 2005–2006 Hotbar.com started sending cease and desist letters to security companies which defined Hotbar as "malware". As a countermeasure, Symantec successfully sued Hotbar.com to get clear right to put it in their "low risk adware"-category. Hotbar can be detected and removed by several anti-spyware and anti-virus programs, including Windows Defender, Spybot - Search & Destroy, Nod 32, and Norton AntiVirus. However, some utilities do not remove Hotbar completely and leave Windows registry keys and files behind. Seekmo Seekmo is an adware program by Zango that claimed to be a free tool to provide content such as mp3 files, screen savers, and videos. Seekmo can pop up advertisements even when users have pop-up blockers on their computers, and monitors computer usage to generate ads that users are more likely to respond to. In the process, the program can consume processing power and network bandwidth, slowing down the computer and interrupting other programs. Upon downloading a Seekmo-containing file, license acquisition will begin and the program (Windows Media Player, for example) will display a window describing the content and what Seekmo is. Upon clicking "Accept", it will automatically send a file for download called "Setup.exe". After downloading this file users must install the Seekmo Toolbar, which will display advertisements related to the websites visited. Seekmo is still available, and identifies itself as belonging to Pinball Corp.; operation may no longer be as described above. Zango Messenger Zango Messenger is an adware-containing adaptation of the instant messaging client Easy Message. Originally developed by Ryan Dewsbury, it was acquired by Zango in February 2005. It connects to the AIM, Yahoo! Messenger, MSN Messenger and ICQ networks. References External links Adware Firm Driven Under by Security Industry – PC Magazine Removal of Zango www3.ca.com Companies based in Bellevue, Washington Software companies of the United States Software companies of Israel Defunct software companies Companies disestablished in 2009
25210971
https://en.wikipedia.org/wiki/Component%20Developer%20Magazine
Component Developer Magazine
Component Developer Magazine or CoDe (registered as CODE Magazine) is a computer magazine edited and produced by the publishing and software company EPS Software. CODE is published bi-monthly, and it is available in printed and digital format. The magazine focuses on providing information about Computer programming in using today's development topics including .NET Framework, Visual Studio, Microsoft SQL Server and Microsoft SharePoint. History CoDe Magazine was started in 1999, by Rick Strahl and EPS Software’s President Markus Egger. The first three issues of the magazine were written in German, and originally published in Austria (then called Software Developer Magazine). Shortly after the original three issues were published, .NET technologies became a primary focus for the magazine as Microsoft was beginning to phase out Visual FoxPro. In 1999, CoDe Magazine’s parent company, EPS Software, moved its headquarters to Houston, where it is still located. EPS Software and CODE participate in community developer events and frequently support Code Camps, or code and development camps put on by the community to help promote software development. See also List of computer magazines References CoDe magazine at WorldCat.org External links CODE magazine official website. 1999 establishments in Austria Bimonthly magazines published in the United States Computer magazines published in the United States German-language magazines Magazines established in 1999 Magazines published in Texas Mass media in Houston
5141199
https://en.wikipedia.org/wiki/NeoPaint
NeoPaint
{{Infobox software | name = NeoPaint | logo = Nphd.png | screenshot = Neowin.PNG | caption = NeoPaint for Windows | developer = OSCS Software Development, Inc.(1991-1993)NeoSoft Corp (1993 - present) | released = | latest release version = 5.3.0 | latest release date = | latest preview version = | latest preview date = | programming language = | operating_system = MS-DOS,<ref>Subject: Kites & Computers (translation of a DRAma article), Date: Mon, 16 May 1994 06:07:01 -1000, From: Simo Salanne, ...3. Graphics design...then import the sail layout into a paint program like Neopaint II...NeoPaint - Version 2.0...OSCS Software Development, Inc...Bend, OR 97701-4631 U.S.A. NeoPaint $45 + $21 (shipping and handling)</ref> Microsoft Windows | platform = | language = English | genre = Raster graphics editor | license = Proprietary software | website = | size = 10MB }} NeoPaint is a raster graphics editor for Windows and MS-DOS. It supports several file formats including JPEG, GIF, BMP, PNG, and TIFF. The developer, NeoSoft, advertises NeoPaint as "being simple enough for use by children while remaining powerful enough for the purposes of advanced image editing"''. The first version, NeoPaint 1.0, was release in 1992 on two floppy disks. It supported video modes ranging from 640x350 to 1024x768 and multiple fonts. NeoPaint 2.2 came out for MS-DOS 3.1 in 1993, with support of for 2, 16, or 256 color images in Hercules, EGA, VGA, and Super VGA modes. NeoPaint 3.1 was released in 1995 supporting 24-bit images and formats like PCX, TIFF and BMP. NeoPaint 3.2 was released in 1996. An updated version, NeoPaint 3.2a, supported the GIF file format. NeoPaint 3.2d was released in 1998. A Windows 95 version named NeoPaint for Windows v4.0 was released in 1999 supporting the PNG file format. On September 1, 2018 the program was rebranded as PixelNEO becoming one of the VisualNEO software products. Formats such as JPEG 2000, ICO, CUR, PSD and RAW are supported. References External links PixelNEO 1992 software Raster graphics editors Graphics software Graphics Windows graphics-related software DOS software
64538017
https://en.wikipedia.org/wiki/Institute%20of%20Computer%20Engineers%20of%20the%20Philippines
Institute of Computer Engineers of the Philippines
The Institute of Computer Engineers of the Philippines (ICpEP, formerly Philippine Institute of Computer Engineers) is a non-profit professional organization of computer engineers in the Philippines. It is registered under the Securities and Exchange Commission as a non-stock and non-profit organization. ICpEP is also the official computer engineering organization for academic community and industry practitioners in the Philippines. The ICpEP is also the organization that manages and prepares certification board exams for computer engineers in the Philippines, as well as the recognized organization of Certified Computer Engineers (CCpE) and Professional Computer Engineers (PCpE). History In 1992, a group of computer engineers formed a professional organization for computer engineers, the Philippine Institute of Computer Engineers, or PhICEs. But after years of being active and conducting conventions, seminars, and symposia to fellow computer engineers and students, PhICEs became inactive. In 2008, several computer engineers tried to revive the professional organization after it's been inactive for years. Then the organization was reformed as Institute of Computer Engineers of the Philippines. History of ICpEP.SE ICpEP.SE or Institute of Computer Engineers of the Philippines Student Edition, is a student chapter of the ICpEP, which aims to link academics to various colleges and universities. The first group of student chapters started in 2008, it was founded by 11 schools, namely, Adamson University, Asia Pacific College, Central Colleges of the Philippines, De La Salle University, Far Eastern University Institute of Technology, Mapua Institute of Technology, Pamantasan ng Lungsod ng Maynila, Polytechnic University of the Philippines-Manila, STI College-Recto, Technological Institute of the Philippines-Manila, and Technological Institute of the Philippines-Quezon City. And now, there are more than sixty-eight schools across the Philippines that have an ICpEP student chapter. See also Institute of Electronics Engineers of the Philippines Computer engineering References External links ICpEP official website ICpEP Singapore Chapter website ICpEP Facebook page Professional associations based in the Philippines Computer science-related professional associations Engineering societies 1992 establishments in the Philippines