url
stringlengths
15
1.48k
date
timestamp[s]
file_path
stringlengths
125
155
language_score
float64
0.65
1
token_count
int64
75
32.8k
dump
stringclasses
96 values
global_id
stringlengths
41
46
lang
stringclasses
1 value
text
stringlengths
295
153k
domain
stringclasses
67 values
https://orgerlab.org/research/
2024-04-19T21:20:56
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817455.17/warc/CC-MAIN-20240419203449-20240419233449-00135.warc.gz
0.912343
1,119
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__3940300
en
Our goal is to understand how the brain integrates sensory information and selects and executes appropriate actions. In particular, we aim to determine the organization and function of neural circuits underlying visually guided behaviors. We use the zebrafish as a model organism because it allows us to visualize and manipulate activity in neural circuits throughout a vertebrate brain. As early as one week post-fertilization, zebrafish display a rich repertoire of innate visual behaviors, following moving patterns, avoiding predators and tracking and capturing live prey. With no skull and transparent skin, the entire volume of the brain can be imaged non-invasively in one field of view, and many neurons are individually identifiable from fish to fish. Determine the principles on which sensorimotor circuits are organised and reveal how activity dynamics unfold throughout the whole brain during behaviour High-speed behaviour tracking, Optogenetics, Whole-brain calcium imaging Models and Regions Zebrafish / Whole-brain The Vision to Action lab uses a combination of advanced optical, genetic and behavioural methods in zebrafish. In recent years, zebrafish have emerged as an attractive model system, as they exhibit a robust set of instinctive visually guided behaviours, while their brain, which follows a typical vertebrate pattern, is sufficiently small and transparent so that researchers can non-invasively image the activity of each of its neurons. Specifically, the team follows the activity of neurons by imaging changes in calcium levels inside neurons, a marker of neural activity, while performing high-speed behaviour tracking to make a detailed, quantitative analysis of visually evoked swimming and eye movements. In addition, the team develops genetic tools in order to probe and manipulate defined circuit elements with high specificity. Recent work from the lab described, at single neuron resolution, the neural activity dynamics throughout the brain of behaving zebrafish. While even very simple behaviours involve activity in hundreds of neurons distributed across many brain regions, the team found that these elaborate patterns are consistent across individuals down to a very fine anatomical spatial scale. In addition, the team has also developed a high-speed, real-time tracking system that has allowed them to systematically characterise the swimming behaviour of zebrafish larvae in response to a variety of different stimuli. Using a computational approach to behaviour classification, called unsupervised machine learning, they have identified a core set of swimming movements, and demonstrated how they are used flexibly across different behaviours. Understanding the Neural Mechanisms that Control Swimming Speed in Zebrafish Larvae Animals often use distinct gaits to move at different speeds, and this requires the engagement of distinct neural circuits. Zebrafish larvae use different motor patterns, and recruit different spinal interneurons, during slow and fast swimming. Currently, it is not known how the brain computes desired speed or relays this information to the spinal cord. We have developed a system to perform high-speed online analysis of tail kinematics in freely swimming fish, while presenting visual stimuli. We find that zebrafish will adjust their swim speed to track different moving patterns, and they do this by switching between two discrete motor patterns. We intend to discover the neural substrates responsible for this behaviour by imaging whole brain neural activity in restrained fish, during visually evoked swimming at different speeds in a closed-loop virtual reality environment. By thoroughly investigating the mechanisms of speed control in zebrafish larvae, from visual inputs to spinal circuits, we hope to uncover general principles of vertebrate locomotor control. Neural circuits underlying the optokinetic response in larval zebrafish How neural circuits integrate sensory information to produce appropriate actions is a fundamental question in neuroscience. We aim to address this question using optokinetic behavior, reflexive eye movements in response to whole field motion. Even these simple responses can involve coordinated activity in hundreds of neurons distributed in areas throughout the brain. We image the pattern of neural activity in the brains of transgenic fish, which express a genetically encoded calcium indicator in all of their neurons, while they track visual stimuli with their eyes. Since this behavior is very repeatable, we can systematically record responses from the whole brain with single cell resolution. Presentation of different stimuli, such as monocular, or binocularly conflicting gratings allows us to determine what sensory or motor signals are represented at each point. These experiments represent the first comprehensive analysis of the neural circuit underlying a sensorimotor behavior in a vertebrate brain. Circuit mechanisms of visuospatial processing in the zebrafish brain Complex visual behaviours, such as capturing moving prey or avoiding approaching predators, require animals to compute the location and salience of different objects moving in 3 dimensions. These computations depend on dynamic interactions between many interconnected visual areas in the brain. We use transgenic expression of optogenetic tools, and in vivo 2-photon functional imaging to reveal the cellular organization of these circuits and the dynamics of visual processing in response to complex stimuli. We aim to: (1) generate driver lines that target gene expression to specific cell types within the fish visual system, (2) characterize visual response properties and functional topography within these populations and (3) analyse the dynamics of population activity in the optic tectum and other visual areas, when the fish is presented with competing visual targets. Using optogenetics and laser ablations we will interfere with defined circuit components, to determine the link between circuit computations and behaviour.
systems_science
https://www.starstandard.org/images/guidelines/DIG2012v1/ch11s03.html
2020-10-24T14:54:54
s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107883636.39/warc/CC-MAIN-20201024135444-20201024165444-00275.warc.gz
0.925679
943
CC-MAIN-2020-45
webtext-fineweb__CC-MAIN-2020-45__0__22293873
en
Standards for Technology in Automotive Retail Malicious software is any software that the user did not authorize to be loaded or software that collects data about a user without their permission. The following is a list of terminology commonly used to describe the various types of malicious software: Spyware- Spyware is any technology that aids in gathering information about a person or organization without their knowledge. On the Internet (where it is sometimes called a Spybot or tracking software), Spyware is programming that is put in someone's computer to secretly gather information about the user and relay it to advertisers or other interested parties. Spyware can get in a computer as a software virus or as the result of installing a new program. Virus- a virus is a program or programming code that replicates by being copied or initiating its copying to another program, computer boot sector or document. Viruses can be transmitted as attachments to an e-mail note or in a downloaded file, or be present on a diskette or CD Worm- a worm is a self-replicating virus that does not alter files but duplicates itself. It is common for worms to be noticed only when their uncontrolled replication consumes system resources, slowing or halting other tasks. Logic bomb- a logic bomb is programming code, inserted surreptitiously or intentionally, that is designed to execute (or "explode") under circumstances such as the lapse of a certain amount of time or the failure of a program user to respond to a program command. It is in effect a delayed-action computer virus or Trojan horse. A logic bomb, when "exploded," may be designed to display or print a spurious message, delete or corrupt data, or have other undesirable effects. Trapdoor- is a method of gaining access to some part of a system other than by the normal procedure (e.g. gaining access without having to supply a password). Hackers who successfully penetrate a system may insert trapdoors to allow them entry at a later date, even if the vulnerability that they originally exploited is closed. There have also been instances of system developers leaving debug trapdoors in software, which are then discovered and exploited by hackers. Trojan (Trojan Horse)- a Trojan horse is a program in which malicious or harmful code is contained inside apparently harmless programming or data in such a way that it can get control and do its chosen form of damage, such as ruining the certain area on your hard disk. A Trojan horse may be widely redistributed as part of a computer virus. RATs (Remote Admin Trojans) - are a special form of Trojan Horse that allows remote control over a machine. These programs are used to steal passwords and other sensitive information. Although they are "invisible", symptoms such as a slow moving system, CD ports opening and closing and unexplained restarting of your computer may manifest. Malware - Malware (for "malicious software") is any program or file that is harmful to a computer user. Thus, malware includes computer viruses, worms, Trojan horses, and also Spyware, programming that gathers information about a computer user without permission. Malicious Font - webpage text that exploits the default method used to de-compress Embedded Open Type Fonts in Windows based programs including Internet Explorer and Outlook. These malicious fonts are designed to trigger a buffer overflow which will disable the security on Windows-based PCs. This allows an intruder to take complete control of the affected computer and remotely execute destructive activities including installing unauthorized programs and manipulating data. Rootkits - Rootkits are a set of software tools used by an intruder to gain and maintain access to a computer system without the user's knowledge. These tools conceal covert running processes, files and system data making them difficult to detect. There are rootkits to penetrate a wide variety of operating systems including Linux, Solaris and versions of Microsoft Windows. A computer with rootkits on it is called a rooted computer. There are three types of rootkits. Below is a description of the characteristics of each: Kernel Rootkits - hide a backdoor on a computer system by using modified code to add or replace a portion of the system's existing kernel code. Usually the new code is added to the kernel via a device driver or loadable module. Kernel rootkits can be especially dangerous because they can be difficult to detect without appropriate software. Library Rootkits - hide information about the intruder by manipulating system calls with patches, hooks, or replacements. Application Rootkits - replace or modify regular application binaries with camouflaged fakes, hooks, patches, or injected code.
systems_science
https://www.bowa.com/is-solar-energy-in-the-future-for-your-home/
2024-04-22T07:46:31
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296818081.81/warc/CC-MAIN-20240422051258-20240422081258-00784.warc.gz
0.946384
1,035
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__90261852
en
As the price of electricity continues to climb, many homeowners are looking for new ways to shave down their energy costs. And, these days, more and more are considering renewable energy sources as an option. Technological advancements and an increase in manufacturing volume have significantly lowered the price of solar energy so it now costs about seven times less than it did twenty years ago. As the cleanest and most abundant renewable energy source available, solar power is definitely worth a closer look to see if it’s right for your home. Solar Electric Power A solar electric system harnesses the sun’s energy through photovoltaics (PV), which is the use of solar cells to directly convert light into electricity at the atomic level. The solar cells can be encapsulated in a variety of forms, including panels that are ground or roof mounted in a low-profile or tilt array; thin films that are applied to a building’s structure; or even integrated, architectural roofing tiles. The units serve as semi-conductors, by capturing the sun’s light and causing the electrons in the material to create an electrical current. At this time, most residential solar electric systems are designed to offset 25-50% of the home’s electric needs due to system expense and/or production restrictions imposed by utility companies. Choosing the right solar configuration for your home depends on the local climate and landscape, how much conventional power you’ll want to offset with solar power, your budget, and the amount of space available. There are three main types of electric systems: - Grid-Connected: Still has utility supplied electricity connected to the property, but it is used only when the solar energy produced isn’t enough to keep up with the home’s demand. A meter reads how much energy has been produced via solar and subtracts that amount from the total bill. - Grid-Connected with Battery Bank: Works similarly, but collects and stores any excess power produced, and then uses the back-up power on cloudy days and at night. One advantage of this system is you can still have power in the event of an outage. - Off-Grid (or Stand Alone): This system is not tied to any utility lines. These systems are expensive, but can often be cheaper than establishing conventional electricity in remote areas – something to keep in mind if you own a cabin, sailboat, or RV. Solar Thermal Energy In today’s homes, solar thermal systems use the sun’s rays to directly heat water and air for domestic use or for pool and spa heating. Perhaps the most common application, solar pool heating uses the existing pool filtration system to pump water through a solar collector that transfers the heat directly to the pool water. Since solar pool heating does not require an entirely new system, it is often the most cost-effective type of solar energy currently available. In general, solar water heating, whether for a pool or domestic use, offers a relatively quick payback period when you consider electric and fuel savings and available incentives. And, with 35-65% of a home’s electricity bill going towards heating water, a combined solar thermal and solar PV solution may be an excellent way to recognize additional savings over a PV-only system. Get Paid for your Renewable Energy While these systems can be costly to install, there are tax credits and other incentives provided by governments and third parties to help you recoup your initial investment faster. At the federal level, the Residential Renewable Energy Tax Credit allows taxpayers to claim a 30% credit for qualified expenditures on solar-electric systems and solar water heaters, and there are also many programs available at the state and local levels. Additionally, Virginia, Maryland and Washington, DC are among those states that support a system called net metering. If your PV system is producing more energy than your home is using at that moment, your local utility company can buy the excess energy from you, essentially causing your meter to spin backwards and reducing your overall electric bill. Finally, there are some turnkey ways for homeowners to sell or auction Solar Renewable Energy Credits (SRECs) to help offset the cost. The Future of Solar There are numerous reasons why more homeowners are considering solar as an option for their homes. Two factors that play a role for many are growing environmental concerns and the rising cost of non-renewable resources such as coal, natural gas and oil. While the initial cost of many solar systems makes them not yet competitive with conventional energy sources, we are seeing that the time between system installation and cost breakeven is shortening. As more technological advancements and improvements are made, the costs of solar materials and installation are expected to continue to decline. And, as non-renewable energy sources become scarcer traditional energy prices are likely to rise even higher. Combined, these factors will help to close the cost gap between solar and conventional energy. So, while solar may not yet be feasible for everyone, it’s certainly becoming a more viable energy option worth considering.
systems_science
https://welle.readthedocs.io/en/latest/ig/problems.html
2020-09-18T13:32:45
s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400187899.11/warc/CC-MAIN-20200918124116-20200918154116-00064.warc.gz
0.924656
631
CC-MAIN-2020-40
webtext-fineweb__CC-MAIN-2020-40__0__54304018
en
What Problems Does Welle Solve¶ Right Person, Right Access, Right Time A typical life-cycle of an employee involves the - When an employee joins a company, she is granted access to Microsoft Exchange and Microsoft Active Directory. Since she is hired as a database administrator, she is also granted access to Microsoft SQL server and MySQL server. - After joining the company for a month, she needs to collaborate with her colleagues for a new project. Thus she requests access to Microsoft SharePoint. Her manager approves her access. - 2 years after she started work, she changes her job scope. She is now a systems administrator. - All her previous accesses to databases and Microsoft SharePoint are revoked. Her access to Microsoft Exchange and Microsoft Active Directory remains for her day-to-day communication. - She is now granted access to Linux and Windows servers. - Finally, she decides she wants a change of environment. She quits. - On her last day, all her accesses are revoked. The above illustration of an employee life-cycle can be automated by Identity Management software. This process is commonly known as Provisioning. In most organizations, on an annual basis, IT auditors request IT departments to provide them with most recent lists of all systems and user accounts. The auditors will flag out an audit fault if any orphan or dormant account is found on any system. Orphan Account is an account belonging to a user who has since left the organization, while a Dormant Account is an account which has not had any login activity for an extended period of time. Therefore, most IT departments will require managers to perform access review on their direct reports prior to annual audit reviews. IT departments might also ask application owners to review all accounts in their applications. Failing to perform user access reviews on a regular basis will place the organization at a higher risk for: - A resigned employee gaining remote access to the network or email system - Segregation of duties issues if an employee moves to a new department, but retains system privileges from the previous department - Misuse of dormant administrative accounts that are still active - System compromise through the use of contractor/vendor passwords that never expire The above illustration of a user access review can be automated by Identity Governance software. This process is commonly known as Governance. Common Pain Points¶ - Ineffective and Inefficient Manual Work Orders - Repeated Process for Multiple Applications - Unable to Meet SLA and Compliance Requirements - Un-Tracked Toxic Access Combinations - Tediously Manual Compiled Access Reviews Welle is the next generation Access Review engine, which helps support enterprise IT security and regulatory compliance. You can manage the identity life-cycle in your organization in a secure and convenient portal. Welle brings the following benefits to our customers: - Automated User Lifecycle Management - Audit and Security - User Self-Service - To identify Orphan or Dormant Accounts - Access Review - Audit and Reports - Segregation of Duties
systems_science
http://simhospital.sonhs.miami.edu/the-building/all-floors/index.html
2017-03-25T11:39:07
s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218188924.7/warc/CC-MAIN-20170322212948-00013-ip-10-233-31-227.ec2.internal.warc.gz
0.928508
244
CC-MAIN-2017-13
webtext-fineweb__CC-MAIN-2017-13__0__140886500
en
An essential component of the simulation experience is the debriefing session. The rooms’ design and multimedia features will allow the students to review and reflect on their actions and identify areas for improvement focusing on evidence-based practice. The simulation and debriefing experiences will allow for the students to connect the dots and put theory into practice in a safe learning environment. The control rooms and command center are at the heart of both individual patient care and mass casualty simulation experiences. Here, the instructors, simulation specialists and technicians craft dynamic, interactive, real-life scenarios through the creative use of cutting-edge technology, including a hospital incident command system and a video wall. Using sophisticated audiovisual equipment and a video-capturing system, scenarios can be recorded and shared during debriefing. This access to recorded information on their performance during the simulation encounter allows learners to reflect on their performance and identify strengths and areas for improvement in a safe learning environment. Threaded throughout all the floors and spaces of the Simulation Hospital are telemedicine capabilities and video teleconferencing capabilities, electronic medical records (in all “patient”- care areas), as well as inter-professional education and training capabilities.
systems_science
http://icaisc2020.icaisc.eu/InvitedTalks
2023-09-22T23:52:48
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233506429.78/warc/CC-MAIN-20230922234442-20230923024442-00106.warc.gz
0.926531
1,319
CC-MAIN-2023-40
webtext-fineweb__CC-MAIN-2023-40__0__185591665
en
Włodzisław Duch "Neurocognitive Technologies and Computational Intelligence for Human Augmentation" Center for Modern Interdisciplinary Technologies, Nicolaus Copernicus University, Poland Google: W. Duch, CV: http://www.is.umk.pl/~duch/cv/cv.html Artificial Intelligence has great impact on every aspect of technology, including neurotechnologies used for human augmentation. In recent years progress in methods of brain activity measurement, analysis of neuroimaging and electrophysiological data, and understanding of brain processes, opens new areas for transdisciplinary applications. Identifying patterns of EEG/MEG, ECoG or fMRI signals that serve as "fingerprints" of high subnetwork activity allows for many applications: linking brain activity with thoughts, intentions, emotions and other mental states, objective diagnostic methods in neuropsychiatry, reliable brain-computer interfaces (BCI), optimization of brain processes through neurofeedback, therapeutic interventions using neuromodulation, neurorehabilitation based on direct brain stimulation combined with behavioral procedures. Some commercial applications for treating epilepsy, major depression and other mental problems are already on the market. Although I will focus on technical aspects of brain fingerprinting it is worth to reflect how neurocognitive technologies will enable human-computer interaction, and in an unprecedented way will change the very nature of people, their social interactions and coupling with physical environment. Tom Gedeon "Predicting human internal states from physiological signals" Tom Gedeon is Chair Professor of Computer Science at the Australian National University. He is formerly Deputy Dean and Head of Computer Science at ANU. His BSc and PhD are from the University of Western Australia, and Grad Dip Management from UNSW. He is twice a former President of the Asia-Pacific Neural Network Assembly, and former President of the Computing Research and Education Association of Australasia. He is currently a member of the Australian Research Council's College of Experts. He is an associate editor of the IEEE Transactions on Fuzzy Systems, and the INNS/Elsevier journal Neural Tom's research focuses on bio-inspired computing (mainly neural, deep learning, fuzzy and evolutionary) and human centred computing (mainly eye gaze, wearable physiological signals, fNIRS, thermal, EEG) to construct truly responsive computer systems (biometrics and affective computing) and humanly useful information resources (hierarchical and time series knowledge), industrial (mining, defence) and social good (medical, educational) applications. Human beings reflect their internal states in many ways in their physiological signals, from skin conductivity, heart rate, pupil dilation, brain signals and behavioural measures. Many of these can be collected unobtrusively. The kinds of internal states we have investigated include stress, depression, emotion veracity, and doubt. We have shown in a number of such areas that physiological signals recorded from a human observer can be used to predict the ground truth in the observed data better than the same human beings themselves can do. That is, by the use of appropriately cross-validated machine learning training, we can access implicit knowledge within the human participants, which is not available to their consciousness. Jarek Gryz "Algorithms and Politics" York University, Toronto, Canada, http://www.cs.yorku.ca/~jarek/ In the last few years, interpretability of classification models has become a very active area of research. Both ACM and IEEE initiated new interdisciplinary conferences where fairness, accountability and transparency of "black-box" algorithms is the main topic. Suddenly, computer programs are being evaluated from moral and political point of view. In this talk, I will discuss a couple of recent controversies in this area. First, I will talk briefly about a supposed racial bias in a COMPAS system, widely used in US courts. Second, I will discuss the concept of algorithm interpretability in a more specific legal context. In 2018 EU introduced General Data Protection Regulation with a Right to Explanation for people subjected to automated decision making. The Regulation itself is very brief on what such a right might imply. I will attempt to explain what the Right to Explanation may involve. I then will argue that this right would be very difficult to implement due to technical challenges. I also maintain that the Right to Explanation may not be needed and sometimes may even be harmful. Bartosz Krawczyk "Learning from imbalanced and difficult data" Bartosz Krawczyk is an assistant professor in the Department of Computer Science, Virginia Commonwealth University, Richmond VA, USA, where he heads the Machine Learning and Stream Mining Lab. He obtained his M.Sc. and Ph.D. degrees from Wroclaw University of Science and Technology, Poland, in 2012 and 2015 respectively. Dr. Krawczyk's current research interests include machine learning, data streams, ensemble learning, class imbalance, and explainable artificial intelligence. He has authored more than 60 journal papers and over 100 contributions to conferences. Dr. Krawczyk has coauthored the book Learning from Imbalanced Data Sets (Springer 2018). He was a recipient of numerous prestigious awards for his scientific achievements such as IEEE Richard Merwin Scholarship, IEEE Outstanding Leadership Award, and Amazon Machine Learning Award among others. He served as a Guest Editor for four journal special issues and as a Chair for fifteen special session and workshops. Dr. Krawczyk is a member of the Program Committee for conferences such as AAAI, IJCAI and IJCNN. He is the member of the editorial board for Applied Soft Computing (Elsevier). Learning from imbalanced data is considered one of the vital challenges in contemporary machine learning. Despite more than three decades of research, the problem of handling skewed distributions is still as important as ever, with new challenges emerging on regular basis. This talk will give an overview of the imbalanced learning domain, focusing on contemporary challenging scenarios and recent developments. Special attention will be given to data-level difficulties and understanding minority classes, multi-class imbalanced problems, and data streams with dynamically evolving classes. The talk will discuss various resampling methods, low-dimensional embeddings, algorithm-level modifications, and ensemble learning approaches that were recently proposed to efficiently handle such challenging scenarios.
systems_science
http://www.globalsecurity.org/military/systems/ship/systems/emals.htm
2017-02-24T21:39:48
s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501171630.65/warc/CC-MAIN-20170219104611-00108-ip-10-171-10-108.ec2.internal.warc.gz
0.938531
4,027
CC-MAIN-2017-09
webtext-fineweb__CC-MAIN-2017-09__0__129185910
en
Electromagnetic Aircraft Launch System - EMALS The U.S. Navy is presently pursuing electromagnetic launch technology to replace the existing steam catapults on current and future aircraft carriers. The steam catapults are large, heavy, and operate without feedback control. They impart large transient loads to the airframe and are difficult and time consuming to maintain. The steam catapult is also approaching its operational limit with the present complement of naval aircraft. The inexorable trend towards heavier, faster aircraft will soon result in launch energy requirements that exceed the capability of the steam catapult. An electromagnetic launch system offers higher launch energy capability, as well as substantial improvements in areas other than performance. These include reduced weight, volume, and maintenance; and increased controllability, availability, reliability, and efficiency. The present EMALS design centers around a linear synchronous motor, supplied power from pulsed disk alternators through a cycloconverter. Average power, obtained from an independent source on the host platform, is stored kinetically in the rotors of the disk alternators. It is then released in a 2-3 second pulse during a launch. This high frequency power is fed to the cycloconverter which acts as a rising voltage, rising frequency source to the launch motor. The linear synchronous motor takes the power from the cycloconverter and accelerates the aircraft down the launch stroke, all the while providing "real time" closed loop control. The average power from the prime power is rectified and then fed to inverters. With power from the inverters, the four disk alternators operate as motors and spin up the rotors in the 45 seconds between launches. The disk alternator is a dual stator, axial field, permanent magnet machine. The rotor serves both as the kinetic energy storage component and the field source during power generation and is sandwiched between the two stators. There are two separate windings in the stators, one for motoring and the other for power generation. The motor windings are placed deeper in the slots for better thermal conduction to the outside casing. The generator windings are closer to the air gap to reduce the reactance during the pulse generation. The use of high strength permanent magnets allows for a high pole pair number, 20, which gives a better utilization of the overall active area. The rotor is an inconel forging with an inconel hoop for prestress. The four disk alternators are mounted in a torque frame and are paired in counter-rotating pairs to reduce the torque and gyroscopic effects. The rotors operate at a maximum of 6400 rpm and store a total of 121 MJ each. This gives an energy density of 18.1 KJ/KG, excluding the torque frame. Each disk alternator is a six phase machine with phase resistance and reactance of 8.6 mW and 10.4 mH, respectively. At max speed, the output of one of the disk alternators would be 81.6 MW into a matched load. The frequency of this output is 2133 Hz and drops to 1735 Hz at the end of the pulse, for a max launch. Machine excitation is provided by the NdBFe 35 MGOe permanent magnets, which are housed in the rotor. These magnets have a residual induction of 1.05 T at 40 oC and create an average working air gap flux density of 0.976 T, with tooth flux densities approaching 1.7 T. The stator consists of a radially slotted laminated core with 240 active slots and liquid cold plate. The maximum back EMF developed is 1122 V. Maximum output voltage is 1700 V (L-L) peak and current is 6400 A peak per phase. The disk alternator's overall efficiency is 89.3%, with total losses of 127 KW per alternator. This heat transfers out of the disk alternator through a cold plate on the outside of each stator. The coolant is a WEG mixture with a flow rate of 151 liters/minute. The average temperature of the copper is 84oC, while the back iron temperature is 61oC. The cycloconverter, or power electronics in general, is the pivotal technology allowing EMALS to become a reality aboard ship. With a 103 m long motor, power electronics permit efficient operation by turning on only the coils that can affect the launch at a particular time rather than the entire motor at once. It also permits EMALS to operate at its most efficient point at all speeds by allowing for a variable voltage, variable frequency supply. The cycloconverter is a naturally commutated 3f-1f bridge circuit. The output of one bridge is paralleled/seriesed with outputs of other bridges to attain the power levels required. By paralleling/seriesing the bridge outputs and not the switches themselves, the design eliminates the current sharing reactors and the series capacitors. The output of a cyclo is 0-644 Hz and 0-1520 V(L-L). Simulations of the operation of the cycloconverter have been completed. The peak current output waveform of the cycloconverter is 6400 A for a max launch. The cooling for the switching assembles takes place through liquid cold plates to which the components are mounted. The medium is de-ionized water at 35oC input, 100 psig max, 1363 liters/minute. This is required to dissipate 528 KW lost in the cycloconverters. The launch motor is a linear synchronous "coilgun". The trough is the same as the steam catapult trough to allow for backfit capability. The motor itself is a dual, vertical stator configuration with the active area facing outwards. The rotor, or carriage, sits over the stators much like a saddle and protrudes through the flight deck to be attached to the aircraft. The carriage contains 160 full permanent magnets, the same type used in the disk alternator, NdBFe. The carriage is restrained in two axes by rollers. The rollers run in channels welded to the stator frame. This allows both the stator and trough to flex with the ship and the carriage to follow this flexure while maintaining a consistent air gap of 6.35 mm. The stator consists of 0.640 m long segments, which are 0.686 m high and almost 0.076 m thick. These segments turn on and off as the carriage passes. The position sense system is based on Hall Effect sensors, much as in today's rotary brushless commutated motors. The stators are protected by offsetting them from the slot in the flight deck. This is due to the contaminants, typically jet fuel, nuts, bolts, wrenches, hydraulic oil, etc., that constantly invade the trough through the slot and could, over time, affect the stators. Between the stators, in an environmentally sealed housing, are the busbars and the static switches, which are SCRs used to control the power to the stator segments. The launcher stator is based on the modular unit called a segment. There are a total of 298 segments, 149 per side, for the entire launch motor, each 0.640 m long. The segment is wound as a three phase lap winding with 6 turns per slot and a total of 24 slots. This translates to 8 poles per segment and a pole pitch of 8 cm. These coils are epoxied into a slotless stator structure with G10 separating the coil legs. The slotless stator design keeps the phase inductance low at 18 mH. The phase resistance is 41 mW while the bus resistance is 0.67 mW. The air gap working flux is 0.896 T with the armature reaction of approximately 0.24 T. At full thrust, the permanent magnets experience a shear stress of 38 psi. At the end of the 103 m power stroke, the front of the carriage enters the brake. This brake consists of shorted stator segments, which act as eddy current brakes. At the same point in time, the carriage is still covering a number of active stator segments. Two phases are switched in these segments so that reverse thrust is initiated to help with the braking force. With a projected efficiency of 70% and peak losses of 13.3 MW in the stator, active cooling will be necessary. Maximum coil action is 4.36e6 A2(squared)s, resulting in a maximum copper temperature delta of 118.2 Degrees C. The launch motor has an aluminum cold plate to remove this heat from the attached stator windings and back iron. The cold plates consist of stainless steel tubes in an aluminum casting. The peak temperature reaches approximately 155oC and, after cooling for the 45 second cycle time, cools to 75oC. The carriage that houses the permanent magnets will be cooled by convection, since there will be only slight heating from eddy currents in the carriage structure and magnets. The introduction of EMALS would have an overall positive impact on the ship. The launch engine is capable of a high thrust density, as shown by the half scale model that demonstrated 1322 psi over its cross section. This is compared to the relatively low 450 psi of the steam catapult. The same is true with energy storage devices, which would be analogous to the steam catapult's steam accumulator. The low energy density of the steam accumulator would be replaced by high energy density flywheels. These flywheels provide energy densities of 28 KJ/KG. The increased densities would reduce the system's volume and would allow for more room for vital support equipment on the host platform. Another advantage of EMALS is that it would reduce manning requirements by inspecting and troubleshooting itself. This would be a significant improvement over the present system, which requires substantial manual inspection and maintenance. The EMALS, however, will require a transition of expertise from mechanical to electrical/electronic. EMALS eliminates the complexity of the present system's conglomeration of different subsystems. The steam catapult uses about 614 kg of steam for a launch, it uses hydraulics extensively, water for braking, and electromechanics. These subsystems, along with their associated pumps, motors, and control systems tend to complicate the launch system as a whole. With EMALS, launching, braking, and retraction would be achieved by the launch motor, thereby reducing all the auxiliary components and simplifying the overall system. The hydraulic oils, compressed air, etc. would be eliminated as well as the cylinder lubricating oil that is expelled into the environment with each shot. The EMALS would be a stand alone system, completely independent of the ship's main plant. This will allow greater flexibility in the design of the ship and more efficient ship propulsion schemes. One of the major advantages of electromagnetic launch is the ability to integrate into the all electric ship. The Navy has directed substantial research into its Advanced Surface Machinery program that is developing electric derived propulsion schemes for the next generation of surface combatants. There has also been a good deal of work in high power electric weapon systems -. As such, more and more of a ship's systems will evolve into the electrical counterparts of old mechanical systems. This is true of the launch, and eventually, the arresting gear. The average power required by EMALS is only 6.35 MVA. Taking these power levels off the grid should not be a problem in an all electric ship, considering multimegawatt pumps already exist on carriers for various applications. Perhaps the most interesting aspect of electromagnetic launch is the flexibility it offers in the way of future aircraft and ship designs. An electromagnetic launcher could easily be sized down to perform as a launch-assist system, augmenting the short takeoff of a STOVL aircraft. It can also be easily incorporated into the contour of a ramp, which provides a more efficient fly-away angle for the aircraft being launched. This reduces the required endspeed, the commensurate energy supplied, as well as the stresses on the airframe. Overall, an EM launcher offers a great deal of flexibility to future naval requirements and ship designs. On the other hand, there are drawbacks to the EMALS. One of these is that high power electromagnetic motors create electromagnetic interference (EMI) with electronic equipment. As in the case of an electromagnetic launcher, there would be sensitive aircraft equipment sitting directly above the launch motor. Along with the aircraft equipment is the ship's own equipment, which may be affected by the electromagnetic emissions. Through proper EMC design and a "magnetically closed" motor design, EMI will be minimized. Another drawback of an electromagnetic launcher is the high speed rotating machinery associated with pulsed power applications. The disk alternator rotors are spinning at 6400 rpm, each storing 121 MJ, for a total of 484 MJ. In a laboratory, this is not a problem, but put these rotors on a heaving, jarring platform and it becomes more complicated. In order to ensure safe operation, the flywheel and bearings are to be a stiffer design than conventional. Due to the inherent high level of elegant control of electronic equipment, it is possible to reduce the stresses imparted to the aircraft. The present steam catapult has relatively high peak-tomean acceleration profiles (nominally 1.25, with excursions up to 2.0). This results in high stresses in the airframe and generally poor performance. With an electromagnetic system it would be possible to correct for deviations in the acceleration profile in typically hundreds of milliseconds, which would result in low peak-tomeans. A simulation was conducted that analyzed the level of controllability of the proposed design. The acceleration profile is smooth and flat, compared with a typical steam catapult profile. The simulation shows that for various load conditions, the EMALS is capable of operating within the 1.05 max peak-to-mean acceleration requirement. The result of this reduced peak-to-mean is reduced stress on the airframe. To quantify the effects of a reduced peak-to-mean, a Fracture Mechanics analysis was conducted on the airframe with both the steam catapult and EMALS peak-to-means. The results from this analysis show a peak airframe life extension of 31% due to the reduced stresses on the airframe. This is becoming more important as tight budgets are forcing the Navy to procure fewer aircraft. This also has the benefit of a safer operational environment, since when the EMALS experiences any unforeseen problems during a launch, it has the capability to quickly adjust and correct for them, even if a component fails during the launch. The EMALS offers the increased energy capability necessary to launch the next generation of carrier based aircraft. The steam catapult is presently operating near its design limit of approximately 95 MJ. The EMALS has a delivered energy capability of 122 MJ, a 29% increase. This will provide a means of launching all present naval carrier based aircraft and those in the foreseeable future. The program is now in DEMVAL in a Critical Component Demonstration (CCD) phase. It is fundamentally a risk reduction phase, in which components, subsystems, and systems which pose the greatest amount of technical risk will be researched and developed to ensure that the technical issues are manageable before proceeding to full scale design. The components being developed are the cycloconverter, the stator, permanent magnets, and control system. These components are required to individually demonstrate their full design capability. For the cycloconverter this means power density, waveform generation, thermal management, and all the per unit electrical parameters the design requires. For the stator section, thrust density, thermal management, and all the per unit design parameters must be demonstrated. The permanent magnets must be able to withstand the harsh environment of the present catapult trough. This includes heat, cold, corrosive agents, shock, etc. Once these components demonstrate their design requirements, they will be integrated with each other in a test fixture. This complete launch test fixture will enable the components to mimic a launch system. It will verify the operation of the EMALS, at all speeds and thrust levels, to the required specifications of the overall launch system. Also, electromagnetic interference (EMI) is an issue that must be addressed early on in the design process. It must be fully understood and manageable before proceeding to the next phases of development. The high fields occur in relative close proximity to the aircraft, which houses sensitive avionics, weapons, and magnetic anomaly detection gear. It is, therefore, of prime importance to ascertain the probability of EMI between the EMALS and its neighboring systems. CCD offers a chance to address the issue of EMC. Using an electromagnetic FEA code, the shielding effectiveness of the catapult trough will be determined, as will the effects of various trough geometries. This model will be verified with hardware at low power levels. Once there is good agreement between the simulation and the empirical data, the simulation will be scaled up to the levels of EMALS. The simulation model takes advantage of the symmetry of the trough and launch motor in the Y-axis. Since the major area of concern of the EMI issue is the fields on the flight deck, only the end turns are modeled. The coil legs running in the vertical direction will contribute little to the fields on the deck. This simulation is run at the complete spectrum of frequencies that the launch motor will produce, and the fields above the deck will be compared to the sensitivities of the various aircraft equipments. The magnetic vector potential A for a 100 Hz, 10,000 A source. This is just a representative source to show the shielding effectiveness of the trough. Little energy is escaping the trough structure. The magnetic fields are 0.07 mTat 10 cm above the deck at the center of the slot. Along the flight deck, the fields reach a maximum of 0.3 mT within 2.5 cm above the deck right over the coil. They fall to the Earth's ambient Electromagnetic motors for both launching and recovery of aircraft aboard a carrier are now possible due to a myriad of technical advancements. The advantages of electromagnetic motors are their improved performance capability over present systems and the resultant reduced weight and volume because of the high power, force, and energy densities possible. These savings are especially important on a carrier where they are precious commodities. In the future Navy, weight and volume may be of even higher importance as smaller budgets may demand smaller ships, and future design will require, just as in automobiles and space vehicles, etc., more performance out of smaller boxes. Electromagnetics offers this advantage. These systems would also provide the inherent controllability that comes with electrical machinery allowing for safer, less All prospective EMALS designs must be capable of launching present and future naval fixed wing aircraft from the deck of an aircraft carrier. Design goals for the program are: 30% reduction in manning, 20% reduction in life cycle cost, 20% improvement in operational availability, and up to a 50% reduction in installed size and weight when compared to the current steam catapults. The performance goals for the EMALS are:90,000,000 ft-lbs. of energy capability, an end speed range between 55 and 200 knots, a peak to mean acceleration of 1.05 for all aircraft launches (including light weight aircraft such as Unmanned Air Vehicles and Unmanned Combat Air Vehicles), and a cycle time of 45 seconds. The Navy defines the EMALS as consisting of four major subsystems: an energy storage subsystem, power conditioning subsystem, a launch engine, and a control system. |Join the GlobalSecurity.org mailing list|
systems_science
https://ncce.no/en/arbeidsgrupper/
2023-10-04T10:48:38
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233511364.23/warc/CC-MAIN-20231004084230-20231004114230-00589.warc.gz
0.933332
313
CC-MAIN-2023-40
webtext-fineweb__CC-MAIN-2023-40__0__214004424
en
WHAT WE DO One of the biggest challenges we face in ensuring the future growth of our industry and private sector is the scarcity of resources and the shortage of raw materials. We are currently using natural resources at a much faster rate than nature is able to produce them, while only 20 per cent of waste streams have satisfactory recycling solutions. To tackle this challenge, while also helping to meet our climate commitments, NCCE is working to achieve a faster transition to a circular economy through the increased use of side and waste streams as an input factor for new products, value chains and markets. We aim to do this by facilitating cooperation, innovation, and good operating conditions for our members. Members contribute to this goal by: - Developing new technology and circular business models for waste and side streams. - Developing new value chains and markets for waste and side streams across existing industries. - Influencing operating conditions to open new markets for products based on waste and side streams. We are a practical-minded cluster, and we establish working groups centred around different topics to achieve our goals. Here, members can get involved as participants and be the driving force to solve our collective challenges in collaboration with other cluster members. At present, we have working groups for multiple topics, and the groups themselves organise how they work. The working groups bring a resource from the administration or the board. Feel free to contact us if you would like to participate or if you have input on topics other than the ones we currently have:
systems_science
https://edit.lsa.umich.edu/physics/directory/faculty/norifranco_ci
2016-08-30T10:47:51
s3://commoncrawl/crawl-data/CC-MAIN-2016-36/segments/1471982976017.72/warc/CC-MAIN-20160823200936-00234-ip-10-153-172-175.ec2.internal.warc.gz
0.78643
1,081
CC-MAIN-2016-36
webtext-fineweb__CC-MAIN-2016-36__0__129455317
en
Professor Franco Nori’s research is in theoretical condensed matter physics and quantum information processing. He has also done research in computational physics, transport phenomena (e.g., of vortices or electrons), energy conversion and solar energy, as well as the dynamics of complex systems. His research work is interdisciplinary and also explores the interface between atomic physics, quantum optics, nano-science, and computing. His research group is also studying artificial photosynthesis, light-to-electricity conversion, nano-mechanics, hybrid quantum electro-mechanical systems, quantum nano-electronics and quantum emulators. Particular emphasis is being placed on superconducting Josephson-junction qubits, scalable quantum circuitry and improved designs for their quantum control. An underlying theme of his work is to better understand nano-scale quantum systems and devise methods to control them. His research group uses physical models to make predictions that can be tested experimentally and that can be used to better understand the observed phenomena. Nori is a Fellow of the American Physical Society, the Optical Society of America, the UK’s Institute of Physics, and the American Association for the Advancement of Science (AAAS). In 2014, he received the Matsuo Foundation Award; and, in 2013, the Prize for Science and Technology by the Minister of Education, Culture, Sports, Science and Technology of Japan. Hybrid Quantum Circuits: Superconducting Circuits Interacting with Other Quantum Systems, (Z.-L. Xiang, S. Ashhab, J.Q. You, F. Nori), Rev. Mod. Phys. 85, 623 (2013). Quantum Biology, (N. Lambert, et al.), Nature Physics 9, 10-18 (2013). Photon Trajectories, Anomalous Velocities and Weak Measurements: A Classical Interpretation, (K.Y. Bliokh, A.Y. Bekshaev, A.G. Kofman, F. Nori), New J. Phys. 15, 073022 (2013). Relativistic Hall Effect, (K.Y. Bliokh, F. Nori), Phys. Rev. Lett. 108, 120403 (2012). Stimulating Uncertainty: Amplifying the Quantum Vacuum with Superconducting Circuits, (P.D. Nation, J.R. Johansson, M.P. Blencowe, F. Nori), Rev. Mod. Phys. 84, 1-24 (2012). Atomic Physics and Quantum Optics using Superconducting Circuits, (J.Q. You, F. Nori), Nature 474, 589 (2011). Observation of the Dynamical Casimir Effect in a Superconducting Circuit, (C.M. Wilson, et al.), Nature 479, 376 (2011). Physics World Top Five Breakthroughs of the Year 2011. Also, the top Readers' choice of 2011 on Nature News. Artificial Photosynthetic Reaction Centers Coupled to Light-harvesting Antennas, (P.K. Ghosh, A.Yu. Smirnov, F. Nori), Phys. Rev. E 84, 061138 (2011). Natural and Artificial Atoms for Quantum Computation, (I. Buluta, S. Ashhab, F. Nori), Reports on Progress in Physics 74, 104401 (2011). Relativistic Electron Vortex Beams: Angular Momentum and Spin-orbit Interaction, (K.Y. Bliokh, M.R. Dennis, F. Nori), Phys. Rev. Lett. 107, 174802 (2011). Distinguishing Quantum and Classical Transport through Nanostructures, (N. Lambert, C. Emary, Y. N. Chen, F. Nori), Phys. Rev. Lett. 105, 176801 (2011). Terahertz Josephson Plasma Waves in Layered Superconductors: Spectrum, Generation, Nonlinear, and Quantum Phenomena, (S. Savel'ev, V. A. Yampol'skii, A. L. Rakhmanov, F. Nori), Rep. Prog. Phys. 73, 026501 (2010) Landau-Zener-Stuckelberg Interferometry, (S.N. Shevchenko, S. Ashhab, F. Nori), Phys. Reports 492, 1 (2010). Quantum Simulators, (I. Buluta, F. Nori), Science 326, 108 (2009). The Dynamical Casimir Effect in a Superconducting Coplanar Waveguide, (J. R. Johansson, G. Johansson, C. M. Wilson, F. Nori), Phys. Rev. Lett. 103, 147003 (2009). Modelling Light-driven Proton Pumps in Artificial Photosynthetic Reaction Centers, (P. K. Ghosh, A. Y. Smirnov, F. Nori), J. Chem. Phys. 131, 035102 (2009).
systems_science
https://innovelconsulting.com/sap-concur-integration/
2024-02-27T03:50:17
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474670.19/warc/CC-MAIN-20240227021813-20240227051813-00832.warc.gz
0.917727
219
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__123483645
en
Innovel team has expertise in helping businesses of all sizes and sectors automate processes and sync data between SAP Concur solutions and other ERP business applications. From streamlining processes to syncing data seamlessly and reducing the need for connectors and middleware, we can help you achieve automation, drive efficiency and improve visibility across your operations. With a reliable SAP Concur integration, you can instantly reduce the need for manual data entry and reduce the risk of errors. You can finally combine your business operations into one integrated system by connecting SAP Concur with other SAP modules like SAP ECC and S/4HANA. Whether you need an ERP-integrated travel and expense management system or a streamlined process for document processing, we can provide you with all the tools and resources needed to start unleashing the potential of SAP Concur. By introducing automated processes and synchronised data, you can: Our SAP Concur integration services provide a reliable and secure connection between your SAP and other ERP solutions. What’s the end result? Greater versatility and visibility over your business operations.
systems_science
https://www.einhell.fr/en/services/smart-garden-pump/
2023-12-07T01:25:40
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100626.1/warc/CC-MAIN-20231206230347-20231207020347-00523.warc.gz
0.891263
1,351
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__161922115
en
GE-AW 1144 SMART Automatic home and garden pump The smart domestic water system combines all the benefits in a single device and, thanks to the Einhell Connect app, the GE-AW 1144 SMART can also be integrated with other smart devices. - SMART CONNECTIVITY: Control via Einhell Connect app - Can be controlled from ANYWHERE using our app - Large water filler plug for easy priming of the pump - Water drain plug protects against frost damage; Built-in thermal protection - Dry run protection - Integral pre-filter - Convenient main on/off switch - Flow rate: max. 5,000 L/h - Discharge head: max. 48 m - Suction head: max. 8 m - Max. pressure: 4.8 bar - Pressure connector: 33.3 mm (R1 internal thread) - Suction connector: 42 mm (R1 1/4 external thread) Your benefits at a glance More independence with the GE-AW 1144 SMART This pump is really smart! In Smart Mode, the irrigation system uses artificial intelligence, also known as AI, to learn and adapt to the individual behaviour of the user. The garden pump can use the available data and user behaviour to automatically create schedules that are perfectly tailored to the weather and your preferences. If you give the automatic water pump permission to use your location, it even incorporates data on the current weather. The garden pump learns in which weather conditions and at what time you usually water your garden, and creates its own schedule based on this. This really turns watering the garden into a smart and efficient irrigation process. If you already have other smart devices in your garden, they can automatically coordinate with each other via Smart Mode. You can select which devices are enabled for this beforehand in the Einhell Connect app. No Wi-Fi. No Bluetooth. All you need is a stable cellular connection. This ensures wide coverage and lets you control the device no matter where you are. The smart irrigation system relies on two technologies: GSM (Global System for Mobile Communications) and NB-IoT (NarrowBand Internet of Things). The GSM standard was introduced in the 1990s and is one of the world's leading mobile communications standards. The NB-IoT standard enables devices and equipment to be networked with wide operating reach. This ensures good signal coverage even in places that are hard to access. Even when you're on holiday, you can water your lawn easily via the app. Einhell Connect app The Einhell Connect app turns the Einhell GE-AW 1144 SMART home and garden pump into an automatic irrigation system and brings smart technology to your garden. The app lets you adjust a wide range of settings from the comfort of your sofa. You can control irrigation schedules, view the system status and the time it was last synced, check live readings, such as water temperature or discharge pressure, and be notified of any important messages or information about your irrigation system. Furthermore, other Einhell smart devices can also be integrated and coordinated via the app. You can find more information about the Einhell Connect app here. How to make a Firmware Update Requirements for the USB stick for the firmware update: - The USB stick shouldn´t be larger than 8 GB. - The USB stick shouldn´t contain any hidden files, folders or partitions. - The USB stick must be formatted in FAT32. How to update the Firmware: - Download the file to an empty USB stick. - Before you start installing the firmware update, unplug the pump. - Open the cover and insert the USB stick into the USB port on the back of the pump. - Connect the power plug of the pump to start the update process automatically. You can see that the update process has started when the blue LED of the pump is turned off. - The update process is completed after approx. 1 min. Afterwards remove the USB stick and wait until the blue LED is blinking again. - Once the blue LED lights up permanently, you can check the Firmware version in the App under Settings. - Improved session and reconnection-checks - General improvements Services for your automatic domestic water and garden pump GE-AW 1144 SMART 3-year Einhell product warranty Our complimentary additional service especially for you Register your new Einhell product within the first 30 days of purchase and benefit from the extended warranty. Once you register, we will extend the warranty on your smart pump by an additional year, making a total of 3 years. 3-2-1 and off we go: Find the serial number on your Einhell product and start the registration process! Spare parts service Sustainable, reliable, and the perfect match: Our subsidiary iSC GmbH supplies you and your Einhell products with any spare part you could think of. Even once the statutory warranty period has elapsed, you can be sure that your Einhell products will be professionally maintained and repaired with original components should they become damaged or worn. As a 100% subsidiary of Einhell, iSC GmbH not only offers high-quality original spare parts, but also exemplifies the outstanding service from Einhell: Specialist repair technicians in our in-house workshop take care of each individual repair case to ensure that your smart pump is quickly back in working order and fully operational again. *The support period for firmware updates is at least 2 years from the product launch. Get started quickly and easily Do you have questions about the GE-AW 1144 SMART home water system? We have the answers! #General – Do I need a Wi-Fi or Bluetooth connection to control the pump from the app? No, neither WLAN nor Bluetooth is necessary for controlling the pump via the app. The pump communicates directly with your smartphone via the mobile network. There are no additional costs. #General – Can the pump only be controlled from the app? No, the pump can also be used without the app as a conventional automatic domestic water system. #General – Does the pump have an automatic mode so that it switches on automatically when water is required? Yes, the pump has an integrated pressure switch, which automatically switches the device on and off. #General – Do I need a separate pre-filter to protect the pump from contaminants in the water? No, the pump already has a pre-filter installed.
systems_science
https://www.jordyalblas.nl/en/portfolio/virtue-solar-decathlon
2023-01-30T04:38:34
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499801.40/warc/CC-MAIN-20230130034805-20230130064805-00476.warc.gz
0.97485
528
CC-MAIN-2023-06
webtext-fineweb__CC-MAIN-2023-06__0__293585097
en
We built an app as part of Team Virtue, a university team that built a sustainable house for the Solar Decathlon competition in Dubai. The app helps inhabitants of our house to live more sustainably, both environmentally and socially. Through iterations, we focused our designs on educating users about behavior and lifestyle. The app also showed our vision to the competition's jury and visitors whom all responded very positively. Designer at Team Virtue UX, Interactions, Animations, Front-end dev Team: 2 designers When I joined the team, one team member had already created a basic version of the app. We discussed how we could develop the app in such a way that it would demonstrate the vision of our team to people. We wanted to take the application beyond a regular smart home app and focused on educating people about sustainable behavior and lifestyle. We also wanted to keep the app simple and have a wow-effect: it's part of a competition after all. We made quick sketches as a starting point. House functioning How can we make sure that people really understand the functioning of their house? Education How can we positively educate people about their sustainable behavior and influence their lifestyles? Social LINQ focuses on social interaction and shared facilities. How can we build this into the app? Control over appliances How can we let people have control over appliances and show monitored energy consumption? We used research conducted by previous team members that focused on personalised feedback to influence sustainable behaviour. Besides, we looked at how other apps show (sustainable) data and appliance controls and we analyzed how other apps inform users and motivate behavior change. We developed our app using ReactJS to be able to scale it to multiple platforms. The API and database layer, necessary to communicate with appliances and sensors, was built by another team. We had bi-weekly meetings to discuss architecture and functionality. In the meantime, our team tested the real house on the university campus. It was temporarily opened for visitors and we put iPads in the house with a basic prototype of the app based on mock-up data. With the competition deadline coming up, these weeks taught us to set limitations and constraints. We went to Dubai were the competition was being held and continued developing the app. Each university team had 15 days to build their house and connect their systems. Then, the competition phase started and our house was opened for public and juried tours. During the first days, we were able to fine-tune the application. For instance, we noticed that buttons on the main screen were not clicked often during tours. We redesigned this part.
systems_science
https://www.mistelektronik.com/services/
2022-05-20T04:43:05
s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662531352.50/warc/CC-MAIN-20220520030533-20220520060533-00613.warc.gz
0.914184
473
CC-MAIN-2022-21
webtext-fineweb__CC-MAIN-2022-21__0__58683160
en
We provide hardware design services on a wide range from PowerPC, ARM and Intel based microprocessor to PIC and ARM Cortex-M microcontrollers. In our electronic board design we utilize electronic processor modules to shorten the design process considerably. We serve various industrial sectors such as defense, medical and automation with our electronic hardware design services. We provide embedded software development services on a wide range of architectures such as ARM Cortex-M and PIC based microcontrollers and PowerPC and ARM Cortex-A based microprocessors. We take part in every stage of an embedded software development process. We develop BSPs, embedded Linux applications, QT based GUIs and real-time applications. With the help of our design partners and engineering team we provide rapid design solutions for industrial sectors suchs medical, defense and automation. Production ready prototypes and related design details for mechanic, electronic hardware and embedded software design are prepared and delivered switfly with coordination of our system engineers. As TQ Group Turkey partner we provide pre-sale product consultancy as well as hardware and software support oriented towards after-sale and development of the processor modules. With the starter kits, sample schematics and software that we provide we ensure our customers a rapid product development process. We provide hardware design and embedded software consultancy services in several sectors such as medical devices, defense systems, automation systems etc. on every production development stage. We also provide biomedical systems and biomedical calibration consultancy services with our experienced staff. We provide embedded software development services using the cross platform QT framework. With the power of C++ and QT Framework we work with our customers to design, develop and support applications that are high performance, visiually appealing and platform independent. We provide support of Yocto Build System for TQ-Group Processor Modules. With our hand-crafted system, you can use yocto build system with TQ modules without any hassles. The support services includes generating yocto layers, kernel and BSP porting for your custom boards. We provide support and development services for OpenCL and OpenCL Embedded on TQ Processor Modules. With our extensive experience, you can use OpenCL technology with TQ modules without any hassles. Our services include integration of OpenCL on your platforms and development of algorithms.
systems_science
https://www.newser.com/story/125665/ibm-designer-mark-dean-pcs-headed-way-of-the-typewriter.html
2021-07-25T15:50:04
s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046151699.95/warc/CC-MAIN-20210725143345-20210725173345-00640.warc.gz
0.943962
240
CC-MAIN-2021-31
webtext-fineweb__CC-MAIN-2021-31__0__138640533
en
PCs are losing their place at the forefront of personal computing, says a designer of the first IBM PC. To commemorate the 30th anniversary of that computer, the 5150, Mark Dean writes in a blog post that PCs are “going the way of the vacuum tube, typewriter, vinyl records, CRT and incandescent light bulbs." They’ve helped create a world that’s ripe for new technologies, but their day has passed. And for all our excitement over smartphones and tablets, PCs aren't actually being "replaced at the center of computing" by another gadget, "but by new ideas about the role that computing can play in progress." “These days, it’s becoming clear that innovation flourishes best not on devices but in the social spaces between them, where people and ideas meet and interact. It is there that computing can have the most powerful impact on economy, society and people’s lives.” But don't expect PCs to vanish entirely: “While PCs will continue to be much-used devices, they’re no longer at the leading edge of computing.” (Read more PCs stories.)
systems_science
https://podcast.hpi.de/73-new-episode
2023-12-04T14:02:23
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100529.8/warc/CC-MAIN-20231204115419-20231204145419-00352.warc.gz
0.913705
129
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__12557270
en
The clean-IT Conference is an international platform for exchanging ideas to make the digital world more sustainable. And you can be part of it - on the 25th and 26th of October 2023. Find all information to register and the agenda here. Prof. Ralf Herbrich, HPI Managing Director and chair of the Artificial Intelligence and Sustainability research group will host the clean-IT conference on AI and Sustainability at the Hasso Plattner Institute. In this podcast episode, he talks about how AI models can support the fight against climate change and what needs to be done to reduce the carbon footprint of Artificial Intelligence.
systems_science
http://nightskyinfocus.com/2012/07/
2013-06-19T09:25:14
s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368708546926/warc/CC-MAIN-20130516124906-00075-ip-10-60-113-184.ec2.internal.warc.gz
0.915223
877
CC-MAIN-2013-20
webtext-fineweb__CC-MAIN-2013-20__0__40635525
en
Astrophotographers who regularly travel to remote observing sites require a reliable power source to last an overnight imaging session. In this article, I will describe how to construct a DIY field battery — the most essential component of any portable imaging setup. I will also discuss how to calculate the total power requirement of your system in order to determine the recommended battery capacity (ampere-hours) that will provide continuous power that will last overnight (longer than 12 hours). A DIY field battery to power my DSLR camera, mount, laptop, and dew heaters Passive anti-dew system (i.e., requiring no power like a lens/corrector hood) may be able to help but only up to a certain extent (it delays the formation of dew, perhaps for a few hours, but will not completely stop it). As soon as the temperature of the scope drops below that of the dew point, dew will start to form and you will have no choice but to end your observation early. Without an active anti-dew system (heating by using electricity), it is simply impossible to completely eliminate dew. In this article, I will describe how an inexpensive DIY dew heater could be constructed using nichrome wire as the heating element. DIY dew heater using nichrome wire About a year ago (November 2011), I started constructing a home-built autoguider, a setup astrophotographers use in imaging galaxies, nebula, and many other deep-space stuff. The setup is no different from what is used by observatories world wide, except that this one was built entirely from scratch. Feel free to browse the details of the project here. A home-built autoguider setup showing the key components: (1) imaging telescope, (2) imaging camera, (3) guidescope, (4) guide camera, (5) tracking mount, and (6) a computer. This part of the DIY guide focuses on the actual guiding operation and the drift-alignment method for precise polar alignment. We begin by first assembling the telescope along with the guidescope. We also attach the imaging and the guiding cameras and connect all the necessary cables leading to and from the computer. Screenshot during actual guiding operation Now that we have already devised a simple contraption that allows a computer to convert guiding commands into light pulses, our next task is to devise a way for a telescope mount to “read” these pulses and translate it into actual east-west movement. This part of the DIY guide will describe the wiring diagrams that will enable any computer to talk to any type of telescope mount (i.e., with or without an autoguider port). Wiring diagram of a DIY autoguider Guiding commands from the computer are sent through a port called ‘LPT1‘, or the parallel port (or sometimes called printer port). It is a kind of interface that allows a simple way for a computer to communicate with other devices. We will try to convert these ‘commands‘ into a form that can be easily interpreted by your telescope mount. The simplest way to do that is to convert the commands into light pulses using Light-Emitting Diodes (or LEDs). These light pulses in turn will be used to drive what is called a ‘light activated switch‘ that we will connect directly to the autoguider port or hand controller. In this DIY guide, we will focus first on how a computer (with the use of the guiding software called GuideMaster) can generate light pulses, by connecting LEDs to the computer’s parallel port. The parallel port is mounted on a socket called DB25F(F stands for ‘female socket’) or DB25M(M stands for ‘male socket’). It has 25 pins (1 to 13 top row, 14 to 25 bottom row). For this project, we are only interested in pins 4, 5, and 25 (other pins will be utilized however in future upgrades). Shown below is a photo of my laptop’s parallel port. A female parallel port (DB25F). Note the location of pins 4,5, and 25 (see arrows).
systems_science
https://www.reaction-grp.com/react-air-view
2022-09-28T19:52:19
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335276.85/warc/CC-MAIN-20220928180732-20220928210732-00042.warc.gz
0.892303
220
CC-MAIN-2022-40
webtext-fineweb__CC-MAIN-2022-40__0__162203826
en
Interconnected Air Monitoring System Governments throughout Europe are mobilising new legislation to combat the spread of pathogens within buildings. Among many factors, carbon dioxide levels are being used to measure the quality of the air. The React-Air View is an interconnected monitoring system which measures: Carbon Dioxide (Co2) Volatile Organic Compounds (VOC) Air Quality Index (AQI) Temperature and Humidity These sensors are extremely accurate with a variance of less than 1%. Advanced Monitoring Technology and Cloud Based Online Systems React-Air View modules are placed around your building and information is sent wirelessly back to our React-Air cloud-based system. By downloading our React-Air mobile app, or using a web-browser, our system alerts the system administrator if levels of either Co2 or VOC’s exceed set limits. Historical information is also stored on our cloud portal, so you are able to prove compliance and deliver confidence that you are meeting your duty of care.
systems_science
https://cogsci.ca/2018/11/08/imagenet-classification-with-deep-convolutional-neural-networks-krizhevsky-sutskever-and-hinton/
2023-05-31T09:31:12
s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224646457.49/warc/CC-MAIN-20230531090221-20230531120221-00620.warc.gz
0.90892
320
CC-MAIN-2023-23
webtext-fineweb__CC-MAIN-2023-23__0__5816592
en
We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0% which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overfitting in the fully-connected layers we employed a recently-developed regularization method called “dropout” that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.
systems_science
https://www.qustodio.com/en/business/
2016-05-27T19:59:12
s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464049277091.27/warc/CC-MAIN-20160524002117-00035-ip-10-185-217-139.ec2.internal.warc.gz
0.837821
434
CC-MAIN-2016-22
webtext-fineweb__CC-MAIN-2016-22__0__77692175
en
- Works on: Stop worrying, start managing Qustodio helps you drive productivity by showing how users in your organization use devices and Internet resources. Qustodio enables you to identify and resolve issues before they become problems and eliminate time wasting and inappropriate use of resources. Qustodio puts you in control Seamlessly manage and monitor all device use across traditional desktop computers, laptops, tablets as well as mobile devices. Thanks to Qustodio’s unique real time technology, device usage updates and configuration changes are instantaneous. Some of the things you can manage with Qustodio: See and manage application usage across all devices. Prevent access to inappropriate content across all browsers. Define productive time limits and internet access. Find your devices in real time and get SOS alerts in an emergency. Calls & Texts Monitor or block calls and SMS texts. Covers all popular platforms Qustodio works on Windows, Android, Mac, iOS, Kindle, and is designed to cover all users and all devices including the ones on the move, giving you complete coverage of your organization. Keep up to date on your organization’s activities through a simple dashboard from any Internet enabled device, whether you’re at the office, on the road, or halfway across the world. More great benefits Qustodio is a complete solution and has a rich feature set that lets you do everything with maximum flexibility and at an affordable price. Qustodio makes technology simple and does all the heavy lifting so that you can focus on what’s most important. Multiple Users & Devices Our smart tool easily adapts to any size organization’s needs. Log in and manage settings from any web-enabled device. All Information is presented in clear easy-to-read charts and graphs. Smart Web Filters Protect users and devices from new and potentially harmful websites. Relax knowing Qustodio monitors all questionable activity. Best Customer Service We’ll help you run smoothly with our professional support.
systems_science
http://navelliergrowth.investorplace.com/portfolio-grader/stock-report.html?t=LDOS
2016-02-12T05:38:33
s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701163438.83/warc/CC-MAIN-20160205193923-00335-ip-10-236-182-209.ec2.internal.warc.gz
0.940467
346
CC-MAIN-2016-07
webtext-fineweb__CC-MAIN-2016-07__0__116603113
en
Leidos Holdings, Inc. provides science and technology solutions in national security, engineering, and health areas in the United States. The company offers systems development, technical, and consulting services to federal, state, and local law enforcement, as well as criminal justice agencies; Critical Insight Solutions that helps make sense of large, disparate data sets to detect fraud, advance medical research, protect brands, track geospatial targets, and optimize energy consumption; information security assessments and public key infrastructure solutions for business-critical information, systems, and Websites; financial solutions to government, commercial, and international clients; geospatial products, services, and solutions for various industries, including agriculture, utility, and defense; process improvement and CMMI support; software development; and applied research and technology, information systems, and modeling and simulation solutions for NASA and the military. It also supports phases of drug development, such as discovery, pre-clinical and clinical development, and operations and compliance; develops technologies, engineer solutions, and provide analytical systems for maritime markets; and offers corporate and individual training services comprising eLearning and training evaluation in cybersecurity, interactive multimedia instruction, national security, process improvement and CMMI support, and terrorism response. In addition, the company design and develops technology products, including customized and standard hardware and software, such as automatic equipment identification technology, sensors, and nondestructive imaging and security instruments comprising aviation, geospatial, ocean and marine systems, safety and security, simulation, software, and transportation solutions. The company was formerly known as SAIC, Inc. and changed its name to Leidos Holdings, Inc. in September 2013. Leidos Holdings, Inc. is headquartered in Reston, Virginia.
systems_science
http://keithjp.blogspot.com/2009/11/grain-new-from-grain-10-2009-small.html
2018-06-21T12:33:07
s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267864148.93/warc/CC-MAIN-20180621114153-20180621134153-00535.warc.gz
0.921823
141
CC-MAIN-2018-26
webtext-fineweb__CC-MAIN-2018-26__0__129338270
en
GRAIN | New from GRAIN | 10-2009 - Small farmers can cool the world: "Soils contain enormous amounts of carbon, mostly in the form of organic matter. Evidence provided in this issue of Seedling shows that industrial agriculture over the last half century has led to the leaching out into the atmosphere of large amounts of this carbon. A coordinated global programme, based on simple farming principles, could gradually put back into the soil the organic matter lost over past decades. Available data shows that within 50 years we could capture at least 450 billion tonnes of carbon dioxide, more than two thirds of the current excess CO2 in the atmosphere." Posted by Susan at 6.11.09
systems_science
https://news.stonybrook.edu/facultystaff/steven-skiena-to-direct-stony-brook-institute-for-ai-driven-discovery-and-innovation/
2024-02-22T08:41:38
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947473735.7/warc/CC-MAIN-20240222061937-20240222091937-00422.warc.gz
0.944096
533
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__59337387
en
Distinguished Teaching Professor Steven Skiena, from the Department of Computer Science in Stony Brook University’s College of Engineering and Applied Sciences, has been named Director of the newly formed Institute for AI-Driven Discovery and Innovation. “The advent of AI is ushering in a new era in the history of humanity. Its consequences can be so vast and far reaching that are hard to even imagine today,” said Fotis Sotiropoulos, Dean of the College of Engineering and Applied Sciences. “Stony Brook is well positioned to lead both in advancing AI-centric research and economic development as well as in educating the engineering workforce for the era of intelligent machines. I am really pleased that a computer scientist of the caliber of Professor Skiena will spearhead our efforts in this area.” The Institute will serve as a hub for the AI research effort at Stony Brook University centered around the overarching vision of Human-Machine Symbiosis, based on the idea that AI technology should amplify human intelligence instead of replacing it. For example, by leveraging their respective strengths to compensate each other’s weaknesses, the human-machine partnership becomes mutually beneficial and far more potent at problem solving than what either can do in isolation. With funding from the SUNY Empire Innovation Program, the State of New York and private fundraising efforts, the Institute will focus on four grand challenges — health care, infrastructure, education, and finance — and five foundational research areas — automated and scalable knowledge acquisition, predictive intelligence, explainable AI, trustworthy AI, and ethical AI. In addition to funded research, the Institute will develop and facilitate new educational programs aimed at fueling the workforce for the AI-driven economy of the future. Programs will range from core technical programs to science communication, technology policy and entrepreneurship programs. The Institute will also stimulate the regional and national economy by providing industry and entrepreneurs with advanced technology training and highly skilled graduates. “Recent advances in artificial intelligence and machine learning are dramatically changing the notions of what computers can do,” said Skiena. “I see the Institute as a way to advance research and education in these areas at Stony Brook in Computer Science, Engineering, Medicine and throughout the University.” Skiena, a former Fulbright Scholar, joined the faculty in 1988. His research interests include data science, natural language processing and sentiment analysis, algorithms, and computational biology. Skiena is the author of more than 150 technical papers and six books, including his popular text The Algorithm Design Manual, which is widely used within the tech industry for job interview preparation.
systems_science
https://axelle.me/2022/04/19/rollercoaster-sort-the-best-of-the-two-worlds/
2023-06-07T16:19:41
s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224653930.47/warc/CC-MAIN-20230607143116-20230607173116-00298.warc.gz
0.866857
3,028
CC-MAIN-2023-23
webtext-fineweb__CC-MAIN-2023-23__0__162206202
en
Rollercoaster sort: the best of the two worlds Disclaimer: This article is part of a serie of articles about the Rust Voracious Radix Sort Crate. This article is about the Rollercoaster sort inside the Voracious Radix Sort Crate, not the crate itself. For each article, the prerequisite is to read the previous articles and to understand them. - What is radixsort ? - Voracious sort - Diverting LSD sort - Rollercoaster sort: The best of the two worlds - Peeka sort: Multithreaded radix sort – TBA – When we talk about radix sort, there are two teams. LSD radix sort team vs MSD radix sort team. One can argue that LSD radix sort are faster because all the histograms can be computed at once whereas MSD radix sort is recursive so it is possible to multithread it and it is O(1) space complexity, LSD radix sort is O(n) space complexity. They are both right. So let’s take the best of the two worlds, the LSD radix sort speed and the MSD radix sort recursivity while keeping the O(1) space complexity. The downside is that this sort is very efficient with very large arrays, but performs poorly with small arrays. After intensive testing, I noticed that this sort is very efficient for f64 primitive types. Unless someone tells me that there is a faster sort for these primitive types, I really think this sort is the state of the art for f64 primitive types. For signed integers, there is a heuristic to avoid converting the signed integer to an unsigned integer. This conversion is done because, actually, radix sorts can only sort unsigned integers. When radix sorts sort other types than unsigned integers, there is, under the hood, a conversion from the other type into the unsigned integers. This sort is called « rollercoaster » because it starts with a MSD radix sort approach and then uses a DLSD radix sort approach and then, eventually, it uses again a MSD radix sort approach. So far, I talk only about unstable radix sort. There is a stable radix sort in the Voracious radix sort crate, but I haven’t worked a lot on it. It is just a « naive » version of a stable radix sort. It is still a lot faster than the standard stable Rust Rollercoaster sort theory What would be great if we could compute all the necessary historgrams at once and take into account only the minimum number of bits as it is done in the DLSD sort while using only O(1) space complexity. It would be even better if the sub array could fit inside the CPU cache and that we don’t need to convert signed integers into unsigned integers. What about the no free lunch theorem ? Sadly it still applies. Rollercoaster sort is O(1) space complexity, but there is a big constant. Several histograms can be computed at once, but not all of them. We don’t need to convert signed integers into unsigned integers, but we need to do it at least once. Once again it is a tradeoff. What is complicated is to choose the correct tradeoff to get the maximum efficiency. Rollercoaster: The main idea is to start with a MSD radix sort approach, we compute a histogram on the most significant digits and we have buckets boundaries. If the bucket’s size is smaller than a constant experimentally found and since we know that LSD radix sort approach is faster than MSD radix sort approach, we switch to a DLSD sort on each bucket. So we can have several MSD passes and then, we switch to a DLSD radix sort. The DLSD radix sort can fallback on a MSD radix sort, hence the sort’s name. Histograms: We have to compute one histogram for each MSD passes, and then several histograms at once for the DLSD radix sort. We compute less histograms than when we use only a MSD approach. You can notice that if the constant on which we switch to a DLSD radix sort is big, there are not many MSD passes. So the number of pass to compute histograms is almost the same as if there was only a DLSD approach, which computes less histograms than a pure LSD approach. Space complexity: Since the DLSD is triggered only when a certain threshold is reached, it is O(1) space complexity. Depending on the threshold and the CPU cache, it may even fit into CPU cache. The threshold is still big but by an order of magnitude smaller than a DLSD or LSD approach. Type conversion: To sort signed integers i64 with a radix sort, one has to convert them into unsigned integers. This is done by flipping the first bit. All the other bits are left untouched. Since the first MSD pass is done on the highest level, the other remaining level to sort with the DLSD approach does not need to convert remaining bits since nothing is done on them. After the MSD pass, each buckets are well sorted, there is still to sort elements inside each bucket. We can call another radix sort on the remaining level without flipping the first bit. That way, we spare a lot of useless computation. A complicated fallback: When I said there is a DLSD fallback, it is actually a bit more complicated. The idea is still to use only the minimum number of bits but it comes with a cost, insertion sort cost or boundaries search cost and some other implementation details. Sometimes it is more efficient to use a LSD if there is not many passes left to do. Depending on the conditions, the fallback is a LSD radix sort or a DLSD radix sort. Rollercoaster sort: example I like to explain the theory with an example. I think it is easier to understand. This let’s use i8 instead of u8. As usual for the example, let’s choose a radix equals 2. The binary representation is done with the two complement method. let mut array: Vec<i64> = vec![57, -32, -47, 18, 9, 5, -5, -60, 22, -17]; // In binary it is: [00111001, 11100000, 11010001, 00010010, 00001001, // 00000101, 11111011, 11000100, 00010110, 11101111] The first pass is a MSD pass and since it is i8, we have to convert it first by flipping the first bit. let mut array: Vec<i64> = vec![57, -32, -47, 18, 9, 5, -5, -60, 22, -17]; // In binary it is: [00111001, 11100000, 11010001, 00010010, 00001001, // 00000101, 11111011, 11000100, 00010110, 11101111] // Flip the first bit. // With conversion: [10111001, 01100000, 01010001, 10010010, 10001001, // 10000101, 01111011, 01000100, 10010110, 01101111] // After the first MSD pass. assert_eq!(array, vec![-32, -47, -5, -60, -17, 57, 18, 9, 5, 22]) // In binary it is: [01100000, 01010001, 01111011, 01000100, 01101111 // |________________________________________________| // 01****** // 10111001, 10010010, 10001001, 10000101, 10010110] // |________________________________________________| // 10****** As you can notice, the first MSD pass separates negative values from positive values. Once the first MSD pass is done, the remaining bits to sort are not transformed in anyway, so it is possible to cast the i8 into a u8 – or a i32 into a u32 or a i64 into a u64 – and sort will still be valid. Now the first MSD pass is done, let’s say that we have reach the threshold to switch to the DLSD passes. I recall that: nb_passes = ceil(log2(array.len()) / radix) We have to apply this on each bucket – 10 – . For the first bucket nb_passes = ceil(3 / 2) = 2, we need two DLSD passes and for the second bucket nb_passes = ceil(3 / 2) = 2, we need two DLSD passes too. // We start with: assert_eq!(array, vec![-32, -47, -5, -60, -17, 57, 18, 9, 5, 22]) // In binary it is: [11100000, 11010001, 11111011, 11000100, 11101111 // 00111001, 00010010, 00001001, 00000101, 00010110] // After the first DLSD pass on the first bucket 01: assert_eq!(bucket01, vec![-32, -47, -60, -5, -17]); // In binary it is: [11100000, 11010001, 11000100, 11111011, 11101111] // |____________________________||________||________| // ****00** ****10** ****11** // After the first DLSD pass on the second bucket 10: assert_eq!(bucket10, vec![18, 5, 22, 57, 9]); // In binary it is: [00010010, 00000101, 00010110, 00111001, 00001001] // |________||__________________||__________________| // ****00** ****01** ****10** Now the second DLSD pass: // After the second DLSD pass on the first bucket 01: assert_eq!(bucket01, vec![-60, -47, -32, -17, -5]); // In binary it is: [11000100, 11010001, 11100000, 11101111, 11111011] // |________||________||__________________||________| // **00**** **01**** ****10** **11**** // After the second DLSD pass on the second bucket 10: assert_eq!(bucket10, vec![5, 9, 18, 22, 57]); // In binary it is: [00000101, 00001001, 00010010, 00010110, 00111001] // |__________________||__________________||________| // **00**** **01**** **11**** As you can see, after the second DLSD pass, the two buckets are sorted, hence the whole array. The insertion sort will only check that the bucket is well sorted, there won’t be any swap, so it is very fast. assert_eq!(array, vec![-60, -47, -32, -17, -5, 5, 9, 18, 22, 57]); In this case, the required memory to sort the array was equivalent to half of the array – actually it depends on the threshold -. Histograms were computed twice, once for the MSD pass and once for the DLSD passes. We spare one histogram computation. There were three passes – one MSD and two DLSD – which were the minimum to fully sort the array, we spare one pass if we would have used a « naive » LSD approach. On two passes, we haven’t convert the u8. There was another improvement but not seen in this example, for positive values, there might be « empty » level – level with all the bit equal 0 -, which are skipped. This improvement is never done on signed integers. To summarize, what was improved: - Use of CPU cache if possible – it depends on the threshold – - Use less memory – it depends on the threshold – - Less passes on the whole array to compute histograms - Use minimum amount of pass to sort the whole array, as it is with the DLSD approach - Do not convert the signed integers into unsigned integers within the DLSD passes - Skip empty levels for positive values - Skip empty levels for negative values too, but it is very unlikely One question remains, what is the « best » threshold. This is a very complicated question, because I think the best threshold depends on the CPU. In the crate, I set the threshold with respect to the CPU I was using to develop the Voracious radix sort. Since I have a Ryzen 3950x, I set the threshold to 128_000. Which is huge, but the Rollercoaster sort is very efficient with very big arrays. For smaller arrays, it fallbacks on other sorts. All these improvements can be apply with f64. It is less likely to skip level with float, but for float that do not need to be transformed, it is possible to cast them into u64 and use the DLSD approach. This is why the Rollercoaster sort performs very well on signed integers and floats. The downside, is that because the first pass is a MSD pass followed by DLSD passes, the array needs to be very big so that the sort is efficient. Radix sorts are known to perform poorly on small arrays and most of the radix sorts are implemented for unsigned integers only. Radix sorts are not design to sort something else than unsigned integers. With the Rollercoaster sort, signed integers and floats are well taken into account. For unsigned integers, DLSD sort is the fastest among all the radix sorts in the Voracious radix sort crate. This will be done later. There are all the results in the Github repository.
systems_science
https://salvusdetect.com/
2024-04-18T04:06:55
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817187.10/warc/CC-MAIN-20240418030928-20240418060928-00511.warc.gz
0.78867
165
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__201428151
en
Because tomorrow will be different. Advanced detection technology moving humanity forward. Introducing Salvus Detection Technology and the world’s first handheld chemical and biological interferometric detector. The Salvus Story Salvus offers a flexible, easy-to-use platform to detect contaminants or beneficial substances. The underlying know-how leading to the Salvus technology came from sponsored research conducted at the Georgia Tech Research Institute. With a substantial intellectual property portfolio and more patents pending, Salvus detection technology offers a wide range of applications across multiple industries. Markets We Serve Salvus detection technology offers speed, sensitivity and specificity for demanding applications across multiple industries. About CJB Companies The Salvus Advantage for chemical and biological targets. Sign-up for updates from Salvus.
systems_science
http://opencircuitdesign.com/~tim/research/fpaa/mapld/mapld.html
2018-12-17T13:39:09
s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376828507.57/warc/CC-MAIN-20181217113255-20181217135255-00150.warc.gz
0.806352
372
CC-MAIN-2018-51
webtext-fineweb__CC-MAIN-2018-51__0__51966689
en
Analog Module Architecture for Space-Qualified Field-Programmable Mixed-Signal Arrays R. Timothy Edwards, Kim Strohbehn, Steven E. Jaskulek, and Richard Katz Johns Hopkins University Applied Physics Laboratory 11100 Johns Hopkins Road Laurel, MD 20723-6099> Spacecraft require all manner of both digital and analog circuits. Onboard digital systems are constructed almost exclusively from field-programmable gate array (FPGA) circuits providing numerous advantages over discrete design including high integration density, high reliability, fast turn-around design cycle time, lower mass, volume, and power consumption, and lower parts acquisition and flight qualification costs. Analog and mixed-signal circuits perform tasks ranging from housekeeping to signal conditioning and processing. These circuits are painstakingly designed and built using discrete components due to a lack of options for field-programmability. FPAA (Field-Programmable Analog Array) and FPMA (Field-Programmable Mixed-signal Array) parts exist [1 ] but not in radiation-tolerant technology and not necessarily in an architecture optimal for the design of analog circuits for spaceflight applications. This paper outlines an architecture proposed for an FPAA fabricated in an existing commercial digital CMOS process used to make radiation-tolerant antifuse-based FPGA devices. The primary concerns are the impact of the technology and the overall array architecture on the flexibility of programming, the bandwidth available for high-speed analog circuits, and the accuracy of the components for high-performance applications. R. Timothy Edwards Last updated: October 13, 1999 at 3:11pm
systems_science
https://stockalicious.com/building-blocks-of-performance-specs-of-minecraft-realm-servers/
2023-09-21T18:44:00
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233506029.42/warc/CC-MAIN-20230921174008-20230921204008-00274.warc.gz
0.919623
767
CC-MAIN-2023-40
webtext-fineweb__CC-MAIN-2023-40__0__257611483
en
Building Blocks of Performance: Specs of Minecraft Realm Servers Minecraft, the beloved sandbox game, has captivated millions of players around the world with its endless possibilities for creativity and exploration. Within the Minecraft universe, Realm servers provide a platform for players to connect, collaborate, and embark on shared adventures. To ensure a smooth and enjoyable gameplay experience, the performance of Minecraft Realm servers relies heavily on the underlying hardware and specifications that power them. Understanding the building blocks of performance is crucial for server administrators and players alike. The key components that determine the performance of a Minecraft Realm server include the processor (CPU), memory (RAM), storage, and network capabilities. Each of these elements plays a vital role in delivering a seamless and responsive gameplay experience for all participants. The CPU is the brain of the server, responsible for processing the game logic and managing player interactions. Minecraft is a computationally intensive game, with each block and entity requiring calculations to determine their behavior and interactions. Servers with high-performance CPUs, such as those with multiple cores and high clock speeds, can handle larger numbers of players and complex game environments more effectively. Memory (RAM) is another critical aspect of Minecraft Realm server performance. RAM is responsible for storing temporary data and allowing the server to quickly access information during gameplay. More RAM enables the server to handle a larger number of players, support more simultaneous activities, and reduce latency. Minecraft servers generally benefit from having a sufficient amount of RAM, especially when hosting multiple players or resource-intensive mods and plugins. Storage plays a crucial role in Minecraft Realm servers, as it affects how quickly data can be retrieved and saved. Servers with solid-state drives (SSDs) offer faster read and write speeds, resulting in quicker world loading times, faster chunk generation, and reduced lag. The size of the storage also impacts the server's capacity to store player data, backups, and resource packs. Having ample storage ensures that the server can accommodate the needs of players without running out of space. Network capabilities are essential for facilitating smooth gameplay and minimizing latency. Minecraft is a multiplayer game, and a stable and fast internet connection is vital for a seamless experience. Servers with high bandwidth and low latency connections can handle multiple player interactions and data transfers more effectively, resulting in a smoother gameplay experience for all participants. In addition to the hardware specifications, the software configuration of a Minecraft Realm server also influences its performance. The server software, such as Paper or Spigot, provides optimizations and enhancements that improve performance and stability. Plugins and mods can also impact performance, and it is crucial to select well-optimized and regularly updated ones to minimize any potential performance bottlenecks. It is worth noting that the number of players and the complexity of the game environment can affect server performance. As the number of players increases, the server must handle a higher volume of player interactions, world updates, and calculations, which may strain its resources. Similarly, complex redstone contraptions, massive builds, and resource-intensive mods can put additional strain on the server's hardware. When choosing or setting up a Minecraft Realm server, it is important to consider the hardware specifications carefully. Depending on the expected player count and gameplay requirements, selecting a server with adequate CPU power, RAM, storage, and network capabilities is crucial to ensure a smooth and enjoyable experience for all participants. Understanding the building blocks of performance for Minecraft Realm servers empowers both administrators and players to make informed decisions. By optimizing the hardware and software configurations, server administrators can provide a reliable and responsive environment for players to unleash their creativity and forge connections. Players, on the other hand, can choose servers that offer the performance needed to accommodate their gameplay preferences and aspirations within the vast and immersive world of Minecraft.
systems_science
https://www.aasurg.org/nominations/finance/entry/27563/
2023-02-04T11:43:52
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764500126.0/warc/CC-MAIN-20230204110651-20230204140651-00863.warc.gz
0.942015
192
CC-MAIN-2023-06
webtext-fineweb__CC-MAIN-2023-06__0__221692593
en
Sanjay Mohanty, Active Member Colon & Rectal Surgery Indiana University School of Medicine I am an Assistant Professor of Surgery at Indiana University and health services researcher and implementation scientist. My work is broadly focused on understanding and improving cancer and surgical care delivery to older adults. At current, with collaborators and the support of an NIH K23 award, I am developing machine learning methods to passively predict postoperative delirium, a common and devastating complication that is difficult to recognize at scale, from routinely collected unstructured and structured health care data. I am also interested in using systems science and the tools of network analysis to better understand and improve cancer care delivery in integrated healthcare systems. Societies like the AAS have been instrumental in laying the foundation for my successes to date. I hope to pay it forward by contributing to the development and maintenance of resources for AAS members interested and/or actively engaged in clinical and health services research.
systems_science
http://www.currentautomation.co.za/kaco/kaco-sensors
2018-10-15T23:42:48
s3://commoncrawl/crawl-data/CC-MAIN-2018-43/segments/1539583509958.44/warc/CC-MAIN-20181015225726-20181016011226-00271.warc.gz
0.85124
231
CC-MAIN-2018-43
webtext-fineweb__CC-MAIN-2018-43__0__25256230
en
If you want the best information you need the best equipment. KACO offers reliable and accurate temperature and irradiance sensors to pair with your monitoring hardware. Ambient Temperature Sensor The PT1000 ambient temperature sensor measures the ambient temperature. The sensor measures the ambient temperature using a PT1000 measuring resistance. Temperatures between –50 and 50°C are converted to voltage values between 0 and 10 V. 3 and 4 String PV Combiner The KACO comBINE combiner solution makes PV design with the blueplanet inverters easier and more flexible than ever before. The solar sensor is used for professional monitoring of a photovoltaic system. The measured irradiance can be used to determine the expected yield of a photovoltaic system, which can then be compared with the actual yield. Module Temperature Sensor The module temperature sensor is used by professional PV plant operators to know the temperature of the modules installed in the array. The module temperature sensor converts this reading into a voltage signal. This signal is sent back to the monitoring device. This information is then displayed via the proLOG or the blueplanet professional web portal.
systems_science
https://genie-bpc.github.io/genieBPC/reference/set_synapse_credentials.html
2023-03-25T20:52:19
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296945372.38/warc/CC-MAIN-20230325191930-20230325221930-00114.warc.gz
0.770169
376
CC-MAIN-2023-14
webtext-fineweb__CC-MAIN-2023-14__0__246272180
en
Connect to 'Synapse' APISource: This function sets 'Synapse' credentials for the user's current session. 'Synapse' username. If NULL, package will search environmental variables for `SYNAPSE_USERNAME`. 'Synapse' password. If NULL, package will search environmental variables for `SYNAPSE_PASSWORD`. If your credentials are stored as environmental variables, you do not need to call `set_synapse_credentials()` explicitly each session. To store authentication information in your environmental variables, add the following to your .Renviron file, then restart your R session (tip: you can use `usethis::edit_r_environ()` to easily open/edit this file): `SYNAPSE_USERNAME = <your-username>` `SYNAPSE_PASSWORD = <your-password>` Alternatively, you can pass your username and password to each individual data pull function if preferred, although it is recommended that you manage your passwords outside of your scripts for security purposes.
systems_science
https://articleslister.org/green-light-for-global-greenhouse-gas-tracking-network-global-issues/
2024-04-16T18:10:34
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817103.42/warc/CC-MAIN-20240416155952-20240416185952-00508.warc.gz
0.926101
490
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__124664272
en
The landmark decision comes as heat-trapping greenhouse gas concentrations are at record levels – “higher than at any time over the last 800,000 years”, WMO warned. Data from Earth and space The new Global Greenhouse Gas Watch will combine observations from Earth and from space with modelling, to fill critical information gaps. It will build on WMO’s experience in coordinating international collaboration on weather prediction. The agency said that the exchange of data will be “free and unrestricted”, in support of the Paris Agreement on climate change. According to WMO, between 1990 and 2021, the warming effect on our climate from the main greenhouse gases, carbon dioxide, methane and nitrous oxide, rose by nearly 50 per cent. “We know from our measurements that greenhouse gas concentrations are at record levels”, said WMO Secretary-General Petteri Taalas. “The increase in carbon dioxide levels from 2020 to 2021 was higher than the average growth rate over the past decade and methane saw the biggest year-on-year jump since measurements started. Plenty still to learn “But there are still uncertainties, especially regarding the role in the carbon cycle of the ocean, the land biosphere and the permafrost areas,” said Mr. Taalas. “We therefore need to undertake greenhouse gas monitoring within an integrated Earth System framework in order to be able to account for natural sources and sinks, both as they currently operate and as they will change as a result of a changing climate. This will provide vital information and support for implementation of the Paris Agreement,” he said. Lars Peter Riishojgaard, WMO Deputy Director for infrastructure, said the UN weather agency’s “decision on the generational challenge of climate change mitigation, is a historic step. “Internationally coordinated global greenhouse gas monitoring open to all and operating under WMO’s policy of free and unrestricted exchange of data, will provide valuable, timely and authoritative information on greenhouse gas fluxes to the UNFCCC Parties (the UN climate change convention Secretariat), which will help them in their efforts to mitigate climate change”, he added. Mr. Riishojgaard said there was “very strong support from the science community and private sector”, for the new monitoring project.
systems_science
https://anvir.biblprog.com/
2023-12-02T21:11:44
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100452.79/warc/CC-MAIN-20231202203800-20231202233800-00737.warc.gz
0.861381
650
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__258668408
en
AnVir Task Manager is a free program to manage startup, running processes, services and drivers on a Windows PC. This software is an excellent replacement for the system task manager, which will help you significantly speed up your PC and detect and eliminate some malware. AnVir Task Manager displays complete information about processes, including command line, loaded DLLs, incoming and outgoing traffic, disk usage, tray icon, internet connections, windows, threads, open files, etc. To enhance the protection of AnVir Task Manager offers a security rating for active processes, services, and autoloads. It is based on the program’s behaviour, code analysis and data from the program database. It is also possible to check any suspicious process or file on virustotal.com. The program allows you to view and manage startup programs and services. An integrated database with a description of more than 100,000 programs, services, drivers, and network connections will help in this. Disable or remove programs you don’t need, and your computer will run much faster. And if some program wants to be added to startup, you will receive a notification. You can enable the delayed loading function for some programs to ensure a quick start of the system. To enhance the functionality of Windows, AnVir Task Manager can integrate additional buttons into each window’s title, allowing you to hide the window in a small floating icon, hide the window in the system tray, and dock the window on top of other windows. Also, in the system tray, you can place icons informing about the use of the processor, the temperature of some computer nodes, network traffic, battery charge, etc. AnVir Task Manager is an excellent product that will become an indispensable assistant for Windows PC users. It provides a wealth of information about applications and processes to help you identify and optimize those taking up valuable system resources. At first glance, the program’s interface may seem confusing, but give the program a chance, and you won’t understand how you did without it before. |Updated:||August 6, 2021| |Available languages:||English, French, German, Italian, Polish, Portuguese, Russian, Spanish…| Below are the links to download AnVir Task Manager for Windows for free. In addition to the latest version, there may be links to download previous versions. Choose the distribution you need and download it. All links are direct and checked for malicious inclusions. - Download AnVir Task Manager Free 9.4 exe (4,28 MB) [Windows XP+] - Download AnVir Task Manager (RUS) 9.4 exe (4,24 MB) [Windows XP+] - Download AnVir Task Manager Portable (RUS) 9.4 zip (5,28 MB) [Windows XP+] - Download AnVir Task Manager Free Portable 9.4 zip (5,28 MB) [Windows XP+] - Updated equipment monitoring module and startup program database Download the installer, run it and follow the instructions. At the first launch, familiarize yourself with the program guide.
systems_science
https://cassiefindlay.com/tag/blockchain/
2021-06-21T15:22:02
s3://commoncrawl/crawl-data/CC-MAIN-2021-25/segments/1623488286726.71/warc/CC-MAIN-20210621151134-20210621181134-00203.warc.gz
0.936203
287
CC-MAIN-2021-25
webtext-fineweb__CC-MAIN-2021-25__0__208041752
en
As published recently by the Australian Society of Archivists, in Archives and Manuscripts. Here’s my pre-print. Abstract Recordkeeping professionals build and manage systems that support the creation and maintenance of trustworthy records, however our approach to the design and implementation of such systems has suffered from a lack of innovation and a failure to collaborate … More Participatory cultures, trust technologies and decentralisation: Innovation opportunities for recordkeeping Recordkeeping professionals seek to design systems that will ensure that trustworthy evidence can be relied upon by the communities we serve, and community memory protected over time. However current recordkeeping implementations are flawed, and permit imperfect recordkeeping, inequitable access to records and records loss. Recordkeeping systems implementations have failed to keep pace with trends towards decentralization and personally controlled personal data. The emergence of decentralised trust through computation as seen with blockchain technologies allows us to imagine new models for recordkeeping that can also bring greater assurance of longevity and availability for records users, and offer new opportunities for individuals to keep their own records. In this post, I consider the problem of enabling children in out of home care to make and keep their own records in light of innovations in decentralised trust mechanisms and specifically blockchain technology, and use the core recordkeeping skill of appraisal to better understand how blockchain technologies might form part of the solution. … More Ideas on using appraisal techniques to plan for the use of blockchain technologies in recordkeeping systems
systems_science
https://www.xnetvisibility.com/?page_id=217
2022-12-01T00:39:09
s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446710777.20/warc/CC-MAIN-20221130225142-20221201015142-00229.warc.gz
0.854193
181
CC-MAIN-2022-49
webtext-fineweb__CC-MAIN-2022-49__0__244029070
en
Software-defined storage (SDS) provides the flexible storage foundation you need for hybrid cloud, digital transformation and more. Make business changes as easily as you update your software. Transform your existing infrastructure to meet new challenges. Support traditional and new-generation applications simultaneously with the same infrastructure. Meet the demands of your data-driven infrastructure with a complete storage software family featuring AI-infused capabilities. Software-defined storage software Manage data growth and prepare for the cloud with a bundled suite of SDS solutions. Bring enterprise data services to container environments with a flexible SDS solution for hybrid cloud. Enable AI workloads and consolidate primary and secondary big data storage with industry-leading, cost-effective object storage. More easily deploy fast, highly scalable storage for AI and big data using simple building blocks.
systems_science
https://wearedace.org/dace-sis-is-name-of-our-new-student-information-system/
2019-08-18T10:41:29
s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027313803.9/warc/CC-MAIN-20190818104019-20190818130019-00210.warc.gz
0.931551
252
CC-MAIN-2019-35
webtext-fineweb__CC-MAIN-2019-35__0__160523586
en
Beginning July 1st, 2019 the Division of Adult and Career Education (DACE) will be using an innovative new attendance system called DACE-SIS. This system will marry the functionalities of the current SIS system with state of the art functionalities needed to support students along their unique educational pathway including personalized counseling services. DACE Administrator Rosario Galvan is very excited about the roll out of the new system. “Our goal is to improve functionality by reducing paperwork and to create an efficient student centered registration process. DACE-SIS will also allow us to digitally monitor the Individualized Student Plans ensuring that all students receive counseling and supportive services.” Central Office Advisor, Carlos Rodriguez, who is the DACE-SIS subject matter expert states that, “the new system will include functionalities to optimize efficiencies in tracking innovative new programming such as Integrated Educational Training (IET), Pre-apprenticeship, and the Family Success Initiative. It will also enable adult schools to market and outreach directly to adult learners.” DACE is excited to transition to this new student information system to better remove barriers and change lives! For information regarding DACE-SIS training, click here. Share this Post
systems_science
https://immunology.washington.edu/joan-m-goverman-phd
2018-11-18T02:38:11
s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039743960.44/warc/CC-MAIN-20181118011216-20181118033216-00510.warc.gz
0.88139
864
CC-MAIN-2018-47
webtext-fineweb__CC-MAIN-2018-47__0__24685634
en
- About Immunology - Graduate Program - Postdoctoral Studies - Seminars & Events Joan M. Goverman, Ph.D. Joan M. Goverman, Ph.D. Professor and Chair, Immunology Dr. Goverman received a Bachelor’s degree in Chemistry from Brandeis University. She received a Ph.D. in Biological Chemistry from the University of California, Los Angeles in 1981. A postdoctoral fellowship at UCLA was followed by additional training at the California Institute of Technology. A member of the Department of Molecular Biotechnology when it was founded at the University of Washington in 1992, Dr. Goverman joined the Department of Immunology in 1994. Department of Immunology The Goverman Lab's research focuses on furthering our understanding of the pathogenesis of multiple sclerosis and identifying potential points of therapeutic intervention. Our primary area of interest is delineating mechanisms that lead to the development of multiple sclerosis. Multiple sclerosis is believed to be an autoimmune disease in which self-reactive T cells that recognize myelin proteins escape normal mechanisms of immune tolerance and become activated, causing inflammation and destruction of myelin in the central nervous system (CNS). A large part of our work employs animal models of this disease in order to investigate immune system function within the central nervous system that is not accessible for study in humans. Using animal models, we study the mechanisms that normally maintain immune tolerance to myelin proteins, the triggers that break this tolerance, the characteristics of T cells that mediate the disease and the effector mechanisms that ultimately cause tissue damage. We have developed a number of new animal models using T cell receptor transgenic mice that express T cell receptors from both CD4+ or CD8+ T cells that recognize different myelin proteins. These models recapitulate different aspects of the complex pathology seen in MS patients, and have allowed us to discover novel mechanisms of maintaining tolerance and causing CNS tissue damage. Using these tools, we are investigating how CD4+ and CD8+ T cells and B cells simultaneously contribute to autoimmunity in the CNS. We are defining the parameters that govern T cell trafficking in the CNS, the effector mechanisms that propagate disease and how inflammatory responses are regulated in different microenvironments in the CNS. Our laboratory also uses samples from patients with multiple sclerosis to define the properties of T cells associated with disease. We focus on defining different in pathogenic pathways in multiple sclerosis with the goal of identifying possible points of therapeutic intervention. - Pierson ER, Stromnes IM, Goverman JM. 2014. B cells promote induction of experimental autoimmune encephalomyelitis by facilitating reactivation of T cells in the central nervous system. J. Immunol. Feb 1; 192(3): 929-39 - Simmons SB, Pierson ER, Lee SY, Goverman JM. 2013. Modeling the heterogeneity of multiple sclerosis in animals. Trends Immunol. Aug;34(8):410-22. - Lee SY and Goverman JM. 2013. The influence of T cell Ig mucin-3 signaling on central nervous system autoimmune disease is determined by the effector function of the pathogenic T cells. J Immunol. May 15;190(10):4991-9. - Ji Q, Castelli L and Goverman JM. 2013. MHC class I-restricted myelin epitopes are cross-presented by Tip-DCs that promote determinant spreading to CD8⁺ T cells. Nat Immunol. Mar;14(3):254-61. - Pierson E, Simmons SB, Castelli L and Goverman JM. 2012. Mechanisms regulating regional localization of inflammation during CNS autoimmunity. Immunol Rev. Jul;248(1):205-15. - Ji Q, Perchellet A and Goverman JM. 2010. Viral infection triggers central nervous system autoimmunity via activation of CD8+ T cells expressing dual TCRs. Nat Immunol. Jul;11(7):628-34. B.A., Chemistry, Brandeis University Ph.D., Biological Chemistry, University of California, Los Angeles
systems_science
https://www.lornestewartgroup.com/facilities/what-we-do/ims/
2024-04-21T15:24:23
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817780.88/warc/CC-MAIN-20240421132819-20240421162819-00442.warc.gz
0.959348
412
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__101091553
en
At the heart of our management operations is our very own Information Management System known simply as our iMS. Developed by our in house software team, iMS has been in operation since 2003 and handles all of our planned and reactive tasks across the UK. Unlike most off the shelf CAFM systems, iMS can very quickly be tailored to suit the requirements of any client and being web based it needs only a browser and internet connection to get it up and running. Our customers can be given access to real time job information via a login and password meaning that you can get the information that you need at any time of the day or night on any IT platform or mobile device. Having quality real time data is a key factor in ensuring that jobs are managed from inception to completion and our in house developed, tried and tested iMS software ensures that we have visibility of our mobile and site based activities 24 hours a day. iMS is a great tool for our clients too and is able to provide a wealth of information on demand as well as a suite of reports that can be scheduled to run at a time with a great choice of delivery options giving the data that you need in most of the proprietary formats that are in use today. iMS enables our service centre team to manage multiple job requests across the UK and can be configured to show jobs by latest logged, priority or geography enabling our agents to allocate work efficiently to our field based staff. Every job logged on to the system is allocated a unique ID number that remains with the job right through to invoicing stage ensuring that all jobs can be traced throughout the system. New jobs are allocated a priority based on the contract agreed KPI's with actual job status being updated in real time as jobs are progressed through to completion via the tablets and smart devices carried by our field based teams.
systems_science
https://britishhotelsguide.com/redstone-revolution-engineering-marvels-in-minecraft/
2024-04-22T19:50:14
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296818337.62/warc/CC-MAIN-20240422175900-20240422205900-00190.warc.gz
0.936475
549
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__111325313
en
Minecraft free, a sandbox video game that has captured the hearts of millions, isn’t just about building and surviving; it’s a playground for budding engineers. The introduction of redstone, a mineral that functions as an electrical conductor, has transformed the game into an interactive learning environment where the only limit is one’s imagination. Harnessing the Power of Redstone Redstone can be used to create simple devices like levers and buttons or complex contraptions such as automated farms and intricate roller coasters. This pixelated power source operates much like electricity, allowing players to construct circuits and logic gates, which serve as the building blocks of computational thinking. Young adults can learn the basics of electrical engineering in this immersive world, crafting creations that can do almost anything they can conceive. The Age of Automation One of the marvels of redstone is the ability to automate tasks within the game. Players have developed systems to mine, farm, and even sort items, all without their direct intervention. These automated contraptions showcase the practical applications of the principles of engineering and encourage players to think critically and solve problems in innovative ways. The automation possibilities with redstone are as varied as the imagination, leading to a constant evolution of designs and techniques. Architects of Interactive Entertainment In Minecraft, entertainment takes on a new form with redstone. Ingenious players have built working arcade machines, pianos, and even full-scale computers within the game. These interactive experiences are not just for show; they demonstrate real-world computing concepts such as binary logic and data storage. By delving into these projects, players gain a deeper understanding of the mechanics behind modern technology and the digital world. The Educational Impact of Redstone Engineering Educators have recognized the potential of Minecraft as an engaging learning tool. Redstone engineering serves as a gateway to understanding essential STEM (Science, Technology, Engineering, and Mathematics) concepts in a hands-on and enjoyable manner. By grappling with the challenges of redstone circuitry, young adults develop skills in logic, problem-solving, and design—skills that are highly valuable in today’s technology-driven world. The redstone revolution within Minecraft is more than just a game mechanic; it’s an invitation to explore the realms of electrical engineering and computer science in a uniquely playful context. As players continue to push the boundaries of what’s possible in the game, the influence of Minecraft as an educational tool is likely to grow, inspiring a new generation of engineers and thinkers. Whether you’re a seasoned gamer or a curious novice, the redstone revolution is an adventure in engineering that’s waiting for you to join.
systems_science
https://www.paceruptime.com/pacer/uptime/bankruptcy-watch
2024-04-17T10:18:50
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817146.37/warc/CC-MAIN-20240417075330-20240417105330-00162.warc.gz
0.934904
370
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__36602717
en
BankruptcyWatch helps organizations automate their entire bankruptcy department. We provide a PACER API for easy, electronic access to almost every piece of bankruptcy data found in the PACER system. Gone are the days of manually using the slow and antiquated PACER user interface. BankruptcyWatch clients can save up to 80% of their bankruptcy labor costs by using the BankruptcyWatch PACER API. Using the PACER API, your organization can automate almost all of the bankruptcy data gathering and decisioning process. This means faster, more accurate and better bankruptcy decisions. Many companies manually research 200 data points or more for each case. For example, your bankruptcy staff probably come into the office each day and are presented with a list of cases they need to research and make decisions on. This means a large part of their day is spent researching each case to determine things like income, debts that will survive the bankruptcy, codebtors, affected vehicles, affected real estate, case status, 341 meeting information, post-bk debt to income ratio, etc. Most often, the decisions are clear once all of the data is gathered. BankruptcyWatch can help you automate all of these processes so that when your staff arrive in the morning, they are presented with all the relevant case information along with all of the decisions that were automatically made before they arrived. This frees up more than 3 quarters of your bankruptcy staff's time so they can focus on recovery instead of research. BankruptcyWatch's cutting edge PACER API and case monitoring tools allow you more capacity, better control and less risk than the manual data gathering process. To learn more about how BankruptcyWatch can help you, visit BankruptcyWatch.com, email [email protected] or call 888.895.1328.
systems_science
https://sp3risk.com/2023/01/05/sp3-sitehazards-expanded-to-include-ca-dam-inundation-and-hi-maps/
2024-02-28T09:57:31
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474700.89/warc/CC-MAIN-20240228080245-20240228110245-00246.warc.gz
0.90872
172
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__24245275
en
SP3-SiteHazards Expanded to Include CA Dam Inundation and HI Maps The SP3-SiteHazards software product is regularly updated to include additional data sources as they become available, and our recent updates incorporates recently updated data and maps for Hawaii. These include landslide maps, tsunami maps, site soil information, and prior-event ShakeMaps. SP3-SiteHazards has also been extended to include dam inundation maps in California. These maps show flooding that could result from a hypothetical failure of a dam or its critical appurtenant structure. Simply provide the location of your site, and SP3-SiteHazards will provide a map of dam inundation for each dam that could present a potential hazard to the site. Check out our announcement for more detailed map images and information on these updates.
systems_science
https://math.tongji.edu.cn/info/1385/9845.htm
2024-02-26T16:52:45
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474661.10/warc/CC-MAIN-20240226162136-20240226192136-00373.warc.gz
0.684743
174
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__151533087
en
题目:Some Results on One-Dimension Systems of Quasilinear Wave Equations 报告人:查冬兵 副教授 (东华大学) 摘要:In this talk, for one-dimension systems of quasilinear wave equations with null conditions, we will introduce the following results: 1. Global existence of classical solutions for the Cauchy problem in the small data setting; 2. Global stability of large solutions, including traveling wave solutions and general large solutions with suitable decay property; 3. Global existence of classical solutions for the initial-boundary value problem in the small data setting. Some related unsolved problems will be also discussed.
systems_science
https://forum.apiant.com/t/added-a-conditional-execution-module-in-dev-apiant/289
2022-05-27T21:56:37
s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652663006341.98/warc/CC-MAIN-20220527205437-20220527235437-00402.warc.gz
0.730313
145
CC-MAIN-2022-21
webtext-fineweb__CC-MAIN-2022-21__0__22280816
en
Added a “Conditional Execution” module in dev.apiant that is a hybrid of the existing [Single Value] and [Data Stream] variations. This module allows two scalar values to be tested, while allowing the nested item to emit either a scalar value or a data stream. It fills a gap where the existing [Single Value] module cannot emit a data stream and the existing [Data Stream] module cannot easily test two scalar values without having to place the left side value into a data stream first. An expected common usage scenario is to conditionally execute the nested item based on the output from the When In module. This new module will be in the next system release.
systems_science
https://railsupport.no/en/team/the-roar-andreassen/
2023-12-05T07:47:00
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100550.40/warc/CC-MAIN-20231205073336-20231205103336-00329.warc.gz
0.941362
304
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__255074089
en
Roar Andreassen is a senior advisor in Railsupport and subject manager within RAMS. He has a background in safety and risk management within railways, aviation, emergency networks, the energy sector and telecommunications from Siemens, the Norwegian Civil Aviation Authority, Nokia Siemens Networks and Lloyd's Register Consulting. Roar has experience from design and SW development of safety-critical control systems, design and development of complex society-critical transmission and communication networks as well as more than 10 years of experience with independent investigation and assessment of safety-critical systems within both aviation and railways. Within the railways, Roar has mainly worked with tasks dealing with risk management, availability, reliability and safety. He has several years' experience as a consultant at Bane NOR and Sporveien with a main focus on safety management within ERTMS, signaling systems, traffic management solutions (TMS), railway engineering as well as IT and telecommunications solutions. Roar has a master's degree in data processing and telecommunications and is a college engineer in electronics. Special areas of expertise: •Risk assessment and risk management •Security assessments and technical verifications •European and national railway regulations •RAMS standards EN 50126, EN 50128 and EN 50129 •CSM-RA and technical specifications for interconnection (TSIs) •Independent third-party control (ISA, AsBo, NoBo and DeBo) •Signal technology, ERTMS, train radio, telecommunications and transmission
systems_science
http://davisdesigninc.com/sustainability.html
2021-06-15T21:57:02
s3://commoncrawl/crawl-data/CC-MAIN-2021-25/segments/1623487621627.41/warc/CC-MAIN-20210615211046-20210616001046-00260.warc.gz
0.949786
141
CC-MAIN-2021-25
webtext-fineweb__CC-MAIN-2021-25__0__73310761
en
Davis Design Inc. is now pursuing design and construction based on principles of economic, social, and ecological sustainability. Our aim is to create spaces that reduce environmental impact and reduce the use and need of non-renewable resources. In addition, the implementation of sustainable design features is a significant benefit, as it can result in long term cost savings for our clients. At Davis Design Inc., we not only work at developing sustainability in our work- we also strive to work safely. This commitment to safety is also true for our subcontractors. Consequently, we have worked with industry leading safety professionals to develop and implement a proficient safety system that reflects the interaction between systems, processes and human dynamics.
systems_science
http://www.corepoweredinc.com/blog/page/2/
2018-12-14T20:04:05
s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376826306.47/warc/CC-MAIN-20181214184754-20181214210754-00555.warc.gz
0.952945
4,785
CC-MAIN-2018-51
webtext-fineweb__CC-MAIN-2018-51__0__240014396
en
As we speed along the AI highway, are we getting closer to an ethical cross road? Faced with an automated future, what moral framework should guide us? Image: Matthew Wiebe 21 Oct 2016 Optimizing logistics, detecting fraud, composing art, conducting research, providing translations: intelligent machine systems are transforming our lives for the better. As these systems become more capable, our world becomes more efficient and consequently richer. Tech giants such as Alphabet, Amazon, Facebook, IBM and Microsoft – as well as individuals like Stephen Hawking and Elon Musk – believe that now is the right time to talk about the nearly boundless landscape of artificial intelligence. In many ways, this is just as much a new frontier for ethics and risk assessment as it is for emerging technology. So which issues and conversations keep AI experts up at night? - Unemployment. What happens after the end of jobs? The hierarchy of labour is concerned primarily with automation. As we’ve invented ways to automate jobs, we could create room for people to assume more complex roles, moving from the physical work that dominated the pre-industrial globe to the cognitive labour that characterizes strategic and administrative work in our globalized society. Look at trucking: it currently employs millions of individuals in the United States alone. What will happen to them if the self-driving trucks promised by Tesla’s Elon Musk become widely available in the next decade? But on the other hand, if we consider the lower risk of accidents, self-driving trucks seem like an ethical choice. The same scenario could happen to office workers, as well as to the majority of the workforce in developed countries. This is where we come to the question of how we are going to spend our time. Most people still rely on selling their time to have enough income to sustain themselves and their families. We can only hope that this opportunity will enable people to find meaning in non-labour activities, such as caring for their families, engaging with their communities and learning new ways to contribute to human society. If we succeed with the transition, one day we might look back and think that it was barbaric that human beings were required to sell the majority of their waking time just to be able to live. - Inequality. How do we distribute the wealth created by machines? Our economic system is based on compensation for contribution to the economy, often assessed using an hourly wage. The majority of companies are still dependent on hourly work when it comes to products and services. But by using artificial intelligence, a company can drastically cut down on relying on the human workforce, and this means that revenues will go to fewer people. Consequently, individuals who have ownership in AI-driven companies will make all the money. We are already seeing a widening wealth gap, where start-up founders take home a large portion of the economic surplus they create. In 2014, roughly the same revenues were generated by the three biggest companies in Detroit and the three biggest companies in Silicon Valley … only in Silicon Valley there were 10 times fewer employees. If we’re truly imagining a post-work society, how do we structure a fair post-labour economy? - Humanity. How do machines affect our behaviour and interaction? Artificially intelligent bots are becoming better and better at modelling human conversation and relationships. In 2015, a bot named Eugene Goostman won the Turing Challenge for the first time. In this challenge, human raters used text input to chat with an unknown entity, then guessed whether they had been chatting with a human or a machine. Eugene Goostman fooled more than half of the human raters into thinking they had been talking to a human being. This milestone is only the start of an age where we will frequently interact with machines as if they are humans; whether in customer service or sales. While humans are limited in the attention and kindness that they can expend on another person, artificial bots can channel virtually unlimited resources into building relationships. Even though not many of us are aware of this, we are already witnesses to how machines can trigger the reward centres in the human brain. Just look at click-bait headlines and video games. These headlines are often optimized with A/B testing, a rudimentary form of algorithmic optimization for content to capture our attention. This and other methods are used to make numerous video and mobile games become addictive. Tech addiction is the new frontier of human dependency. On the other hand, maybe we can think of a different use for software, which has already become effective at directing human attention and triggering certain actions. When used right, this could evolve into an opportunity to nudge society towards more beneficial behavior. However, in the wrong hands it could prove detrimental. - Artificial stupidity. How can we guard against mistakes? Intelligence comes from learning, whether you’re human or machine. Systems usually have a training phase in which they “learn” to detect the right patterns and act according to their input. Once a system is fully trained, it can then go into test phase, where it is hit with more examples and we see how it performs. Obviously, the training phase cannot cover all possible examples that a system may deal with in the real world. These systems can be fooled in ways that humans wouldn’t be. For example, random dot patterns can lead a machine to “see” things that aren’t there. If we rely on AI to bring us into a new world of labour, security and efficiency, we need to ensure that the machine performs as planned, and that people can’t overpower it to use it for their own ends. - Racist robots. How do we eliminate AI bias? Though artificial intelligence is capable of a speed and capacity of processing that’s far beyond that of humans, it cannot always be trusted to be fair and neutral. Google and its parent company Alphabet are one of the leaders when it comes to artificial intelligence, as seen in Google’s Photos service, where AI is used to identify people, objects and scenes. But it can go wrong, such as when a camera missed the mark on racial sensitivity, or when a software used to predict future criminals showed bias against black people. We shouldn’t forget that AI systems are created by humans, who can be biased and judgemental. Once again, if used right, or if used by those who strive for social progress, artificial intelligence can become a catalyst for positive change. - Security. How do we keep AI safe from adversaries? The more powerful a technology becomes, the more can it be used for nefarious reasons as well as good. This applies not only to robots produced to replace human soldiers, or autonomous weapons, but to AI systems that can cause damage if used maliciously. Because these fights won’t be fought on the battleground only, cybersecurity will become even more important. After all, we’re dealing with a system that is faster and more capable than us by orders of magnitude. - Evil genies. How do we protect against unintended consequences? It’s not just adversaries we have to worry about. What if artificial intelligence itself turned against us? This doesn’t mean by turning “evil” in the way a human might, or the way AI disasters are depicted in Hollywood movies. Rather, we can imagine an advanced AI system as a “genie in a bottle” that can fulfill wishes, but with terrible unforeseen consequences. In the case of a machine, there is unlikely to be malice at play, only a lack of understanding of the full context in which the wish was made. Imagine an AI system that is asked to eradicate cancer in the world. After a lot of computing, it spits out a formula that does, in fact, bring about the end of cancer – by killing everyone on the planet. The computer would have achieved its goal of “no more cancer” very efficiently, but not in the way humans intended it. - Singularity. How do we stay in control of a complex intelligent system? The reason humans are on top of the food chain is not down to sharp teeth or strong muscles. Human dominance is almost entirely due to our ingenuity and intelligence. We can get the better of bigger, faster, stronger animals because we can create and use tools to control them: both physical tools such as cages and weapons, and cognitive tools like training and conditioning. This poses a serious question about artificial intelligence: will it, one day, have the same advantage over us? We can’t rely on just “pulling the plug” either, because a sufficiently advanced machine may anticipate this move and defend itself. This is what some call the “singularity”: the point in time when human beings are no longer the most intelligent beings on earth. - Robot rights. How do we define the humane treatment of AI? While neuroscientists are still working on unlocking the secrets of conscious experience, we understand more about the basic mechanisms of reward and aversion. We share these mechanisms with even simple animals. In a way, we are building similar mechanisms of reward and aversion in systems of artificial intelligence. For example, reinforcement learning is similar to training a dog: improved performance is reinforced with a virtual reward. Right now, these systems are fairly superficial, but they are becoming more complex and life-like. Could we consider a system to be suffering when its reward functions give it negative input? What’s more, so-called genetic algorithms work by creating many instances of a system at once, of which only the most successful “survive” and combine to form the next generation of instances. This happens over many generations and is a way of improving a system. The unsuccessful instances are deleted. At what point might we consider genetic algorithms a form of mass murder? Once we consider machines as entities that can perceive, feel and act, it’s not a huge leap to ponder their legal status. Should they be treated like animals of comparable intelligence? Will we consider the suffering of “feeling” machines? Some ethical questions are about mitigating suffering, some about risking negative outcomes. While we consider these risks, we should also keep in mind that, on the whole, this technological progress means better lives for everyone. Artificial intelligence has vast potential, and its responsible implementation is up to us. Julia Bossmann, President, Foresight Institute The views expressed in this article are those of the author alone and not the World Economic Forum By 2020, 60% of manufacturers will rely on digital platforms which will support as much as 30% of their overall revenue. IW Staff | Dec 19, 2017 IDC recently released a report, “IDC FutureScape: Worldwide Manufacturing Predictions 2018,” surveying the global manufacturing landscape. When creating its predictions the firm examined ecosystems and experiences, greater intelligence in operational assets and processes, data capitalization, the convergence of information technology (IT) and operations. Most of the group’s predictions refer to a continuum of change and digital transformation (DX) within the wider ecosystem of the manufacturing industry and global economy. “Manufacturers of every size and shape are changing rapidly because of new digital technologies, new competitors, new ecosystems, and new ways of doing business,” said Kimberly Knickle, research vice president, IT Priorities and Strategies, IDC Manufacturing Insights. “Manufacturers that can speed their adoption of digital capabilities in order to create business value will be the leaders of their industry.” Technologies that will have the greatest impact include cloud, mobile, big data and analytics, and internet of things (IoT). Manufacturers also have high expectations for the business value of technologies that are in earlier stages of adoption, such as robotics, cognitive computing/artificial intelligence (AI), 3D printing, augmented reality/virtual reality (AR/VR), and even blockchain. Over the next few years, IDC identified some of the most notable changes in the industry: - Redefining how businesses design (or define), deliver and monetize products and services - Developing new contextualized and customized experiences for customers, employees and partners - Increasing coordination and collaboration between IT and line-of-business organizations, as well as among ecosystem participants - Changing the nature of work and how it’s accomplished with people, process, and technology coming together While the predictions offered largely focus on the near- to midterm (2018–2021), the impact of many of these will be felt for years to come. IDC’s worldwide manufacturing 2018 predictions are: Prediction 1: By 2020, 60% of the top manufacturers will rely on digital platforms that enhance their investments in ecosystems and experiences and support as much as 30% of their overall revenue. Prediction 2: By 2021, 20% of the top manufacturers will depend on a secure backbone of embedded intelligence, using IoT, blockchain, and cognitive, to automate large-scale processes and speed execution times by up to 25%. Prediction 3: By 2020, 75% of all manufacturers will participate in industry clouds, although only one-third of those manufacturers will be monetizing their data contributions. Prediction 4: By 2019, the need to integrate operational technology and information technology as a result of IoT will have led to more than 30% of all IT and OT technical staff having direct project experience in both fields. Prediction 5: By 2019, 50% of manufacturers will be collaborating directly with customers and consumers regarding new and improved product designs through cloud-based crowdsourcing, virtual reality, and product Prediction 6: In 2020, augmented reality and mobile devices will drive the transition to the gig economy in the service industry, with “experts for hire” replacing 20% of dedicated customer and field service workers, starting with consumer durables and electronics. Prediction 7: By the end of 2020, one-third of all manufacturing supply chains will be using analytics-driven cognitive capabilities, thus increasing cost efficiency by 10% and service performance by 5%. Prediction 8: By 2020, 80% of supply chain interactions will happen across cloud-based commerce networks, dramatically improving participants’ resiliency and reducing the impact of supply disruptions by up to one-third. Prediction 9: By 2020, 25% of manufacturers in select subsectors will have balanced production with demand cadence and achieved greater customization through intelligent and flexible assets. Prediction 10: By 2019, 15% of manufacturers that manage data-intensive production and supply chain processes will be leveraging cloud-based execution models that depend on edge analytics to enable real-time visibility and augment operational flexibility. This is good work, a great check and balance, maybe not perfect, however, heading in the right direction. Bias in artificial intelligence can surface in various ways. Some of the best minds are working on the problem. When Timnit Gebru was a student at Stanford University’s prestigious Artificial Intelligence Lab, she ran a project that used Google Street View images of cars to determine the demographic makeup of towns and cities across the U.S. While the AI algorithms did a credible job of predicting income levels and political leanings in a given area, Gebru says her work was susceptible to bias — racial, gender, socio-economic. She was also horrified by a ProPublica report that found a computer program widely used to predict whether a criminal will reoffend discriminated against people of colour. So this year, Gebru, 34, joined a Microsoft Corp. team called FATE — for Fairness, Accountability, Transparency and Ethics in AI. The program was set up three years ago to ferret out biases that creep into AI data and can skew results. “I started to realize that I have to start thinking about things like bias,” says Gebru, who co-founded Black in AI, a group set up to encourage people of colour to join the artificial intelligence field. “Even my own PhD work suffers from whatever issues you’d have with data set bias.” In the popular imagination, the threat from AI tends to the alarmist: self-aware computers turning on their creators and taking over the planet. The reality (at least for now) turns out to be a lot more insidious but no less concerning to the people working in AI labs around the world. Companies, government agencies and hospitals are increasingly turning to machine learning, image recognition and other AI tools to help predict everything from the credit worthiness of a loan applicant to the preferred treatment for a person suffering from cancer. The tools have big blind spots that particularly effect women and minorities. “The worry is if we don’t get this right, we could be making wrong decisions that have critical consequences to someone’s life, health or financial stability,” says Jeannette Wing, director of Columbia University’s Data Sciences Institute. Now, in the wake of several high-profile incidents — including an AI beauty contest that chose predominantly white faces as winners — some of the best minds in the business are working on the bias problem. AI is only as good as the data it learns from. Let’s say programmers are building a computer model to identify dog breeds from images. First, they train the algorithms with photos that are each tagged with breed names. Then they put the program through its paces with untagged photos of Fido and Rover and let the algorithms name the breed based on what they learned from the training data. The programmers see what worked and what didn’t and fine-tune from there. The algorithms continue to learn and improve, and with more time and data are supposed to become more accurate. Unless bias intrudes. Bias can surface in various ways. Sometimes, the training data is insufficiently diverse, prompting the software to guess based on what it “knows.” In 2015, Google’s photo software infamously tagged two Black users “gorillas” because the data lacked enough examples of people of colour. Even when the data accurately mirrors reality, the algorithms still get the answer wrong, incorrectly guessing a particular nurse in a photo or text is female, say, because the data shows fewer men are nurses. In some cases, the algorithms are trained to learn from the people using the software and, over time, pick up the biases of the human users. AI also has a disconcertingly human habit of amplifying stereotypes. PhD students at the University of Virginia and University of Washington examined a public data set of photos and found that the images of people cooking were 33 per cent more likely to picture women than men. When they ran the images through an AI model, the algorithms said women were 68 per cent more likely to appear in the cooking photos. Eliminating bias isn’t easy. Improving the training data is one way. Scientists at Boston University and Microsoft’s New England lab zeroed in on so-called word embeddings — sets of data that serve as a kind of computer dictionary used by all manner of AI programs. In this case, the researchers were looking for gender bias that could lead algorithms to do things such as conclude people named John would make better computer programmers than ones named Mary. In a paper called “Man is to Computer Programmer as Woman is to Homemaker?” the researchers explain how they combed through the data, keeping legitimate correlations (man is to king as woman is to queen, for one) and altering ones that were biased (man is to doctor as woman is to nurse). In doing so, they created a gender-bias-free public data set and are now working on one that removes racial biases. “We have to teach our algorithms which are good associations and which are bad, the same way we teach our kids,” says Adam Kalai, a Microsoft researcher who co-authored the paper. He and researchers including Cynthia Dwork — the academic behind the 2011 push toward AI fairness — have also proposed using different algorithms to classify two groups represented in a set of data, rather than trying to measure everyone with the same yardstick. So for example, female engineering applicants can be evaluated by the criteria best suited to predicting a successful female engineer and not be excluded because they don’t meet criteria that determine success for the larger group. Think of it as algorithmic affirmative action that gets the hiring manager qualified applicants without prejudice or sacrificing too much accuracy. While many researchers work on known problems, Microsoft’s Ece Kamar and Stanford University’s Himabindu Lakkaraju are trying to find black holes in the data. These “unknown unknowns” — a conundrum made famous by former secretary of defence Donald Rumsfeld — are the missing areas in a data set the engineer or researcher doesn’t even realize aren’t there. Using a data set with photos of black dogs and white and brown cats, the software incorrectly labelled a white dog as a cat. Not only was the AI wrong, it was very sure it was right, making it harder to detect the error. Researchers are looking for places where the software had high confidence in its decision, finding mistakes and noting the features that characterize the error. That information is then provided to the system designer with examples they can use to retrain the algorithm. Researchers say it will probably take years to solve the bias problem. While they see promise in various approaches, they consider the challenge not simply technological but legal too because some of the solutions require treating protected classes differently, which isn’t legal everywhere. What’s more, many AI systems are black boxes; the data goes in and the answer comes out without an explanation for the decision. Not only does that make it difficult figuring out how bias creeps in; the opaqueness also makes it hard for the person denied parole or the teacher labelled a low-performer to appeal because they have no way of knowing why that decision was reached. Google researchers are studying how adding some manual restrictions to machine learning systems can make their outputs more understandable without sacrificing output quality, an initiative nicknamed GlassBox. The Defense Advanced Research Projects Agency, or DARPA, is also funding a big effort called explainable AI. The good news is that some of the smartest people in the world have turned their brainpower on the problem. “The field really has woken up and you are seeing some of the best computer scientists, often in concert with social scientists, writing great papers on it,” says University of Washington computer science professor Dan Weld. “There’s been a real call to arms. 5 to 10 years from now we will have a very different manufacturing environment. Specifically due to the exponential technology growth and the impact that Artificial Intelligence and Artificial Consciousness will have on both capability and social impact. This blog is intended to bring awareness of these advances to you and provoke your attention to the social impacts of these developments. I welcome your thoughts, challenges, and perspectives to bring this blog to life and help evolve our social relationship to these changes at the velocity and impact these technologies will have on us.
systems_science
http://teledynecoax.com/pressreleases/2012/CSM_Selection_Guide.asp
2017-08-23T13:45:29
s3://commoncrawl/crawl-data/CC-MAIN-2017-34/segments/1502886120573.0/warc/CC-MAIN-20170823132736-20170823152736-00025.warc.gz
0.82574
705
CC-MAIN-2017-34
webtext-fineweb__CC-MAIN-2017-34__0__187036925
en
Teledyne Coax Switches Announces New Coax Switch Matrix Selection Guide Hawthorne, CA, September 2012: Teledyne Coax Switches introduced its new Coax Switch Matrix Selection Guide designed to feature Teledyne's switch matrix capabilities with comprehensive descriptions of each matrix and additional custom options available. The new 28 page Coax Switch Selection Guide features examples of matrix systems suitable for a variety of markets including: Military and defense, Aircraft, Industrial, SATCOM, Advanced Telecomm, ATE, LTE 4G and many more. Teledyne's 45+ years experience in switching technology make its matrix systems the most comprehensive and reliable on the market today. Teledyne's switch matrix systems are composed of multiple individual switches connected to achieve multi-input and multi-output configurations, allowing users to reduce space and cost. All switching systems utilize Teledyne's universal controller that offers multiple interface options. Teledyne's integrated matrix systems by Teledyne simplify complex switching needs by allowing users to select a combination of input and output ports, instead of tediously commanding individual switches to form a signal path. These matrix systems are available in 50Ω and 75Ω characteristic impendence. Teledyne's Switch Matrix Systems encompass four different series: CSM-1000, CSM-2000, CSM-3000 and CSM-4000. Any matrix system has a variety of options available such as: 1024 switch paths, SMA, mini-SMB, Type N, TNC or 2.92mm standard options, RS-232, USB, GPIB, Parallel TTL, Ethernet TCP/IP interface options, failsafe, latching or normally open configurations, switching systems for 50Ω and 75Ω, internal termination available, 1million life cycle (per port). The CSM-4000 Series is intended to feature non-standard options. This series is flexible and also features RS-232 (Standard), USB, GPIB, Parallel TTL, Ethernet TCP/IP interface options. Some additional options are available such as: bypass, expandable, independent matrices in one chassis, integration of passive components such as filters and attenuators, custom displays, buttons, switches, LEDs and front panel schematics, custom marking, painting, labeling, flanges, handles, non-enclosure switch plates. Other non-standard options include custom matrix interface such as military-rated connectors, indicators, readback, Switching systems for 50Ω & 75Ω applications and an internal termination are available. Customers may order or download the Coax Switch Matrix Selection Guide online at http://www.teledynerelays.com/lit-request.asp. For more information, call (800) 284-7007, visit www.teledynerelays.com or send e-mail to For the latest news releases, visit http://www.teledynerelays.com/newsrelease.asp. For the latest data sheets, visit http://www.teledynerelays.com/newproducts.asp. Teledyne Relays, a Unit of Teledyne Technologies Inc., has been the world's innovative leader in manufacturing ultraminiature, hermetically sealed, electromechanical and solid-state switching products for nearly 50 years. The company's comprehensive product line meets a wide range of requirements for industrial, commercial, military and aerospace uses.
systems_science
https://abcnewstalk.com/nasa-prepares-artemis-i-sls-orion-spacecraft-ahead-of-august-29-launch/
2023-03-23T21:43:48
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296945183.40/warc/CC-MAIN-20230323194025-20230323224025-00146.warc.gz
0.947349
422
CC-MAIN-2023-14
webtext-fineweb__CC-MAIN-2023-14__0__165201155
en
NASA is making ready for Artemis I, the launch of its Area Launch System (SLS) rocket, together with the Orion spacecraft for astronauts, which is about to blast off on August 29. The US house company is readying to check its flight system that’s designed to ship astronauts again to the Moon, a long time after it accomplished its Apollo missions. NASA is making ready for the subsequent technology of house journey. The company’s SLS spacecraft is the newest vertical launch system developed by NASA. Earlier this week, NASA completed a flight readiness evaluate for the Artemis I launch, forward of the scheduled take a look at flight on August 29. The flight directors met at Kennedy Area Heart in Florida and confirmed that the mission was prepared for launch. The SLS-Orion spacecraft is anticipated to blast off on Monday. Artemis I is only the start for NASA, and the company’s take a look at flight is the primary in a collection of more and more complicated missions. The uncrewed flight take a look at will present a basis for human deep house exploration, in keeping with NASA because it plans to return people to the Moon and discover extra of the lunar floor. Final week, NASA’s SLS rocket and the Orion spacecraft for the Artemis I mission arrived on on the Kennedy Area Heart in Florida after a 10-hour journey that started from NASA’s Car Meeting Constructing. Based on NASA, the house company’s engineers and technicians are at present engaged on configuring programs on the pad forward of the launch. The SLS-Orion spacecraft is anticipated to launch on Monday at 8:33am EDT (6:03pm IST). Artemis I’ll stress take a look at the SLS-Orion spacecraft’s programs as a part of NASA’s plans to confirm whether or not the system is able to take astronauts to the moon, a purpose the house company is aiming to finish by 2025, forward of its plans to ship people to different planets, together with Mars.
systems_science
https://www.infologistix.ca/editing-the-hosts-file-on-macos/
2024-03-04T15:53:24
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947476452.25/warc/CC-MAIN-20240304133241-20240304163241-00317.warc.gz
0.764061
241
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__167256583
en
Here’s a trick we use all the time when a client moves to our servers, but the nameservers haven’t been updated yet. This trick lets us get on with work while we wait for DNS propagation. Open Mac Terminal (command line) Applications -> Utilities -> Terminal. In the terminal window you just opened copy/paste the command string below, and press return. sudo nano /private/etc/hosts Edit the hosts file by adding these two lines to the bottom of the file. For example, it should look like: Save save your changes by pressing control-o on your keyboard then return to accept the filename. Exit the editor by pressing control-x. This takes you back to the terminal screen. You may need to flush your Mac’s DNS cache, so copy/paste the below command string into terminal, and press return. sudo dscacheutil -flushcache Pretty much the same on Ubuntu and other Linux flavors, but the hosts file is at etc/hosts. On Windows you can fire up Notepad and open C:\Windows\System32\drivers\etc\hosts
systems_science
https://acipenergy.com/solar-shootout-in-the-san-joaquin-valley/
2023-03-20T10:36:13
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296943471.24/warc/CC-MAIN-20230320083513-20230320113513-00366.warc.gz
0.942308
1,049
CC-MAIN-2023-14
webtext-fineweb__CC-MAIN-2023-14__0__26967753
en
California, United States [RenewableEnergyWorld.com] Side-by-side crystalline and thin-film photovoltaic installations at a water treatment plant in California’s Central Valley should provide a clear indication of which provides the best energy production and cost benefit performance over varying climatic conditions within a year. The 1.6-MW solar array for the South San Joaquin Irrigation District (SSJID) was installed in two phases by Denver-based Conergy Americas in Manteca, CA. Phase One includes 6720 Conergy 175-W crystalline modules mounted on a single-axis solar tracking system that can boost peak-time output by about 15% over fixed systems. The Phase Two tracking system, which went active in late March, uses cadmium telluride (CdTe) modules from First Solar, chosen because they are expected to perform at a lower cost/watt than crystalline modules, according to David Vincent, Western U.S. project director for Conergy. They add 419 kW to the project, and it is believed to be the first commercial thin-film solar tracking system in the U.S. Thin-film modules “can outperform monocrystalline in areas prone to hazy, overcast conditions or in industries that generate dust or high degrees of air particulates,” according to Vincent. They are also superior when there is frequent fog, such as in coastal areas. The reason, he says, is the sensitivity of the thin-film cells to a broader span of the solar spectrum, including infrared and ultraviolet regions. Thin-film cells also should perform better when dust covers the surface, he added. Another advantage of thin-film modules is that less interconnect is needed between cells, so that there is less rise in resistivity and heat loss on hot days, he explained. Early indications, Vincent says, are that the output/DC kW of the thin-film modules is about 10% higher that of the monocrystalline. The project, known as the Robert O. Schultz Solar Farm, will handle almost all of the power needs for a water treatment plant that provides 40 million gallons/day for 155,000 residents and businesses of four nearby communities, as well as irrigation water for 55,000 farm acres. The main goal of the project is to stabilize electrical costs, which can spike in summer months because of time-of-use metering, according to Don Battles, utility systems director for SSJID. Also, these are times when solar output is at a maximum. To reduce long-term maintenance requirements for the thin-film tracking system, the number of drive motors had to be minimized. The challenge was to effectively drive more than 30 tons of modules and steel following the sun’s trajectory with each 2hp motor. This was done by means of a 30-ton screw jack and engineered counter-balance. Power generation data for the crystalline and thin-film modules will be fed from equipment that Conergy installed on inverters to Fat Spaniel Technologies, a nearby monitoring and reporting company. The analysis is put online so that it can be tracked by SSJID’s Battles and his team from offices located more than 20 miles from the solar arrays. The data on the Fat Spaniel Web site also allows the group to compare the 1-MW Phase One solar-tracking system with a number of fixed installations, such as a 1-MW fixed-axis rooftop system at a fruit packing firm in Hanford, CA, a system that Conergy also installed. Battles indicates that the output at the water treatment tracking facility is typically 15%-18% ahead, even though he believes the sun is better at the Hanford location. The irrigation district expects to save nearly $400,000 a year in utility costs due to the solar system, while getting millions of dollars in state incentives. Conergy’s Vincent says that the side-by-side face-off between monocrystalline and thin-film systems is attracting worldwide attention, particularly in Europe where solar has advanced much further than in the U.S. The performance of thin-film modules under the hazy, often foggy conditions is attracting considerable interest in the California valley region, according to Vincent. For example, a 188-kW thin-film fixed solar array is being installed by Conergy in Hanford, CA, for Verdegaal Brothers, a fertilizer, warehousing and soil and water amendment supplier. Vincent said that the First Solar CdTe thin-film installation takes about 10%-15% more ground space, but provides more energy and is expected to cost 10%-15% less than a monocrystalline array. The facility is expected to offset Verdegaal’s utility bills by 99%, cutting some $60,000 a year, while providing for 82% of the company’s energy needs. Over the 25-year life of the system, which is scheduled to start up in July, emissions are expected to be reduced by 6,145 tons of CO2.
systems_science
https://www.familyfirstchiro.net/post/the-autonomic-nervous-system-and-the-chiropractic-adjustments
2024-04-15T09:41:13
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296816954.20/warc/CC-MAIN-20240415080257-20240415110257-00894.warc.gz
0.945546
233
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__66072327
en
Researchers found that chiropractic adjustments have an effect on the Autonomic Nervous System. This is the part of the nervous system that controls things like organ function, metabolic rates, and sleep wake cycles to name a few. Researchers were able to measure the changes in "Edge Light Pupil Cycle Time" (ELPCT) which is one of the light reflexes of the eyes. This reflex is controlled by the Autonomic Nervous System (ANS). A decrease in the Edge Light Pupil Cycle Time was observed after a chiropractic adjustment implying a direct link between a chiropractic adjustment and a response in the Autonomic Nervous System. The study was published in the September 2000 issue of the Journal of Manipulative and Physiologic Therapeutics. So when you hear a chiropractor talk about the immune system, or respiratory rates, or even blood pressure you can see that research like this and other studies like this have confirmed that the chiropractic adjustment does in fact have an impact on the body as a whole. Due to the changes in the central nervous system which controls all other systems of the body.
systems_science
http://robpongi.blogspot.com/2008/10/japanese-plant-midori-san-first.html
2018-07-19T09:55:09
s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676590794.69/warc/CC-MAIN-20180719090301-20180719110301-00469.warc.gz
0.935474
142
CC-MAIN-2018-30
webtext-fineweb__CC-MAIN-2018-30__0__265280192
en
Thursday, October 23, 2008 Japanese Plant 'Midori-san' - First Botanical Blogger Midori-san, which lives on the counter of a Japanese café, writes regular updates with the help of sensors attached to its leaves. The detectors pick up electronic signals on the surface of the plant, which responds to light and human touch. The data is then combined with weather and temperature information and translated into chatty blog posts using a computer algorithm. Midori-san is a hoya kerrii, which are more commonly called "sweetheart plants" because of their heart-shaped leaves. And here is a link to Midori-san's Blog (Japanese)
systems_science
http://simberon.com/faliure_tests.htm
2013-05-22T05:35:03
s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368701370254/warc/CC-MAIN-20130516104930-00061-ip-10-60-113-184.ec2.internal.warc.gz
0.971171
190
CC-MAIN-2013-20
webtext-fineweb__CC-MAIN-2013-20__0__190364738
en
When you're doing test driven development or, in fact, any kind of testing, there's a tendency to test only the success paths of the program to make sure the system does what it's supposed to do. It's just as important, however, to test the failure paths as well. You have to test that the program doesn't do what it's not supposed to do. If your program expects a start date and end date, what happens if the end date is before the start date? What happens if they're equal? If the user is editing an object in a window, what happens if another user deletes that object before the first one saves? If you expect a number from 1 to 10, what happens when an 11 comes in? In my experience, if any bug could possibly occur, it will. It's better for you to find it in the lab than to have the end users find it in the field.Download
systems_science
http://valleyfire.com/services/91-2/
2017-08-17T11:46:45
s3://commoncrawl/crawl-data/CC-MAIN-2017-34/segments/1502886103270.12/warc/CC-MAIN-20170817111816-20170817131816-00295.warc.gz
0.919074
272
CC-MAIN-2017-34
webtext-fineweb__CC-MAIN-2017-34__0__276187624
en
As a result of our long term, mutually beneficial relationship, Siemens – one of the largest electronic and engineering companies in the world – has asked Valley Fire to be one of two Chicagoland distributors for its Sinorix™ Intelligent Fire Suppression Systems. Sinorix, Siemens’ brand name for a product manufactured by DuPont called FM200, is a clean agent (gas) fire suppression system developed in the late 1980s to replace Halon 1301. Compared to Halon 1301, FM200 is less harmful to the environment and non-toxic to humans. There are a number of other FM200 gas systems on the market that are sold under the following trade names: FM200 extinguishes a fire through heat absorption and the release of a small amount of free radicals that inhibit the chain reaction responsible for combustion. It will not damage delicate equipment, which is vital for facilities that cannot afford a fire-related interruption or the loss of data that can result from the activation of a water-based fire sprinkler system. FM200 is stored in cylinders and delivered through a network of piping to distribution nozzles. When smoke detectors detect a fire, the control panel sounds an alarm, shuts down air handlers and disconnects power from the protected equipment. Once these processes have been completed, the system releases the agent for suppression.
systems_science
https://vulcantyre.com/2021/10/28/deciding-on-a-reliable-cloud-data-area-provider/
2022-01-21T17:38:09
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320303512.46/warc/CC-MAIN-20220121162107-20220121192107-00706.warc.gz
0.944513
427
CC-MAIN-2022-05
webtext-fineweb__CC-MAIN-2022-05__0__210672433
en
28 Oct Deciding on a Reliable Cloud Data Area Provider A virtual data area often identifies a organised data center where businesses can retail outlet their electronic and back-up documents, including email boxes, personal pc files, spreadsheets and any other type of digital data. A virtual data room generally serves as a location where firms can risk-free store the organization information they have to conduct organization on a daily basis. These types of service could also be used to provide space for a provider’s archived files and program software. In most instances, a digital data room can be utilized to accomplish the proper research stage during an M&A deal, investment capital investment, or perhaps private equity and mergers and acquisitions. In terms of storage space, digital data areas can vary drastically from a few gigabytes to terabytes, with the ordinary being in regards to hundred gb. The amount of readily available storage uses the volume of user accounts that a particular virtual info room seems to have access to. For example, an organization with hundreds of staff would almost certainly have a storage quantity that is many https://developerszones.com/virtual-data-room-software-2021 terabytes larger than what would be essential by smaller firms. Since the Internet has become this kind of important a part of most businesses today, it is far from surprising that lots of providers nowadays offer impair computing expertise in order to take care of corporate data. However , companies should be aware a number of cloud computing services actually deliver virtual info rooms rather than actual hosting space, and therefore companies must carefully evaluate each company’s demands before choosing a provider. Much like any type of online business service, it is vital that companies be certain that the virtual data place provider they will choose is normally accredited with both HIPAA and EHRP legislation. Furthermore, companies should also ensure that the provider has reviewed their inside IT systems to ensure that each and every one necessary compliance standards happen to be in place. The majority of quality suppliers also have reviews posted on all their websites, making it possible for potential customers to read reviews from other customers.
systems_science
https://agrobiologicalrecords.com/detail.php?view_id=2154
2023-09-22T08:37:58
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233506339.10/warc/CC-MAIN-20230922070214-20230922100214-00284.warc.gz
0.898045
504
CC-MAIN-2023-40
webtext-fineweb__CC-MAIN-2023-40__0__108041664
en
Dua Amna1, Muhammad Rehan Islam1, Ammara Farooq2* and Iqra Munawar3 1Department of Food Science and Technology, Faculty of Food Science and Nutrition, Bahauddin Zakariya University, Multan, Pakistan 2Lahore University of Biological & Applied Sciences (UBAS), Pakistan 3Department of Zoology, Riphah International University, Faisalabad Campus, Pakistan *Corresponding author: [email protected] Bound phenolic compounds (BPs) are abundant in plant-based foods and have gained increasing attention due to their potential health benefits. However, the physiological implications of BPs and their interactions with food matrices remain relatively unexplored. This review aims to provide a comprehensive overview of the functional implications of BPs and the complex interplay between BPs and food components. The digestion process plays a crucial role in determining the bio-accessibility and bioavailability of BPs. Upon reaching the large intestine, BPs encounter the vast and diverse gut microbiota, leading to microbial transformation and the generation of bioactive metabolites. The catabolic activity of gut microbiota, termed colonic fermentation, involves various structural modifications of parent BPs, resulting in the production of microbial metabolites. The characteristics of the food matrix, the type of phenolic–food macromolecule interaction, and the chemical nature of the BPs significantly influence the amount of BPs reaching the colon and the rate of microbial transformation. Moreover, the release kinetics of BPs to the colonic lumen can be modulated through the addition of certain ingredients or technological processes during food processing. The functional implications of BPs on gut microbiota are extensive, including prebiotic effects, anti-microbial properties, and regulation of gut microbial metabolism. BPs can selectively stimulate the growth of beneficial gut bacteria, inhibit the growth of harmful microbes, and influence the production of short-chain fatty acids, which are essential for gut health. Additionally, the synergistic effects of BPs and dietary fiber on gut microbial ecology have been observed, further highlighting the intricate interactions between phenolics and food components. Understanding the functional implications of BPs and their interactions with food matrices is crucial for harnessing their potential health benefits and designing innovative food products with enhanced bioactivity. Further research in this area will shed light on the complex mechanisms underlying BPs' effects on gut microbiota and their overall impact on human health.
systems_science
http://www.ntfsfilerecovery.org/ntfs-data-recovery.html
2017-08-21T21:22:08
s3://commoncrawl/crawl-data/CC-MAIN-2017-34/segments/1502886109670.98/warc/CC-MAIN-20170821211752-20170821231752-00078.warc.gz
0.829675
512
CC-MAIN-2017-34
webtext-fineweb__CC-MAIN-2017-34__0__167552819
en
Are you looking for a Professional Data Recovery Tool for getting back all lost, deleted and damaged files and folders? Are you fed up of finding of so many data recovery software online but not getting the best and result oriented one? Don't worry, We have the finest and truly effective solution for your lost, damaged Windows data. NTFS Data Recovery Software is an impeccable and convenient software to provide with instant Windows Data Recovery Software. It performs data recovery from formatted, re-formatted, deleted Windows NTFS partitions. - Recover lost, corrupted, deleted Windows data files and folders - NTFS data recovery of files emptied from recycle bin - Recover data from corrupted or damaged NTFS file system based partitions - Recover files and folders lost due to corrupt partition table, boot sectors, root folders, or MFT (Master File Table) - Recovery possible in case of accidental formatting of your hard disk, virus attacks - Recovers data even if the disk partition has been formatted with different file system type or with different parameters - Recovers data if you get formatting error - Recovers files with long names - 30-day money back guarantee - Interactive and intuitive graphical interface - Full support for IDE, EIDE, SCSI, SATA, PEN, ZIP drives - Free demo version available NTFS Data Recovery Tool is quick, effortless and effective tool providing you with instant and complete recovery of all lost, deleted, inaccessible folders and files. Embedded with powerful recovery modes namely, Standard Scan and Advance Scan, NTFS Data Recovery Software leaves no stone unturned in retrieving lost files. 'Standard Scan' works well in case of minor data damage, partition deletion, or when files got deleted sometime back. Whereas, 'Advance Scan' proves to be elixir when 'Standard Scan' fails to recover data. It performs NTFS Data Recovery in case of drive formatting, OS reloading, major drive corruptions and damages. NTFS data recovery tool recovers deleted files and folders from NTFS based operating systems, such as Windows 95, Windows 98, Windows ME, Windows NT, Windows 2000, Windows XP, Windows Vista and Windows 7, Windows 8, Windows 8.1, Windows 10, Windows Server 2000, Windows Server 2003, Windows Server 2008, Windows Server 2008R2, Windows Server 2012 installed on the hard disk or any other data storage media like IDE, EIDE, SCSI, SATA, PEN, ZIP drives.
systems_science
https://www.city-cost.com/blogs/City-Cost-News/MgPgj-news
2019-01-20T17:16:08
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583728901.52/warc/CC-MAIN-20190120163942-20190120185942-00563.warc.gz
0.944293
320
CC-MAIN-2019-04
webtext-fineweb__CC-MAIN-2019-04__0__153213475
en
Jan 9, 2018 TOKYO - The government is set to offer information related to disasters to drivers through vehicle navigation systems via Japan's satellite system, a government source said Monday. The government wants to introduce the service that could start as early as fiscal 2018, after many people inside vehicles were not able to obtain crucial information swiftly enough during the 2011 earthquake and tsunami. The system that utilizes Michibiki satellites has an advantage over other communication infrastructure including mobile phone networks as such facilities can be damaged or lost by disaster, the source said. The government is planning to ask companies to cooperate as certain software needs to be installed in vehicle navigation systems, according to the source. The government already conducted demonstration experiments last November in Wakayama and Kochi prefectures, and tested issuing "tsunami warnings" via Michibiki satellites. Separately, the government plans to build a safety confirmation system utilizing the satellites. People can confirm the safety of their family members and friends through the system once evacuees enter their names and other information into the system set up at evacuation shelters. The government plans to introduce the safety confirmation system on a trial basis in five municipal governments during fiscal 2018 and hopes to expand the number to 20 in fiscal 2021, the source said. Kyodo News Plus is an online publication delivering the latest news from Japan. Kyodo News Plus collaborates with City-Cost to bring those stories related to lifestyle and culture to foreigners resident in Japan. For the latest news updates visit the official site at https://english.kyodonews.net
systems_science
https://www.connamara.tech/about-ep3/
2023-03-23T20:03:32
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296945183.40/warc/CC-MAIN-20230323194025-20230323224025-00771.warc.gz
0.918897
867
CC-MAIN-2023-14
webtext-fineweb__CC-MAIN-2023-14__0__283985437
en
EP3: Exchanges Made Easy™ Using advanced technology from trusted market experts, EP3 is the breakthrough exchange platform and order matching engine developed to make it easier to launch a new exchange, improve the performance of a current exchange, or expand into a non-traditional asset class. Conceived by a Trader EP3’s bloodline traces back to the trading pits of Chicago. The platform was conceived by an options trader who started in open outcry pits and has been a pioneer in the transition of trading from the pit to the screen. As a result, a deep understanding of how markets operate is built into EP3 — saving EP3 customers from reinventing the wheel and enabling them to focus on what makes them unique. A Trusted Name in Tech Since 1998 EP3 was incubated within Connamara Systems — a leading capital markets software engineering provider since 1998 — then spun out into Connamara Technologies in 2022. The engineers at Connamara Technologies have deep experience in the rigors necessary to build applications for trading and capital markets. For customers, this means EP3 is built for the scale and reliability that capital markets require. EP3 is the third version of our exchange platform that has evolved over 10+ years. The creators of EP3 took the lessons learned from previous versions and applied them from scratch to build an exchange platform and order matching engine that leverages the scale, reliability, performance, and flexibility afforded by modern software architecture and deployment methods like agile development and cloud deployment. EP3 can be deployed quickly and easily to cloud, hybrid-cloud, and on-premises environments with the use of modern orchestration and containerization methods. This means that EP3 customers can be up and running quickly, focusing on building the business of the exchange, not the exchange technology. Exceptional Professional Support and Services Our Professional Services team can develop solutions around EP3 to address the needs of the market or asset class you are targeting. Our Support Service team is available 24×7 to help keep your exchange running smoothly. Features & Functionality The advanced features and functionality on the EP3 exchange platform and order matching engine are the result of decades of experience in capital markets with the development of flexible, trader-centric tools that improve transparency, liquidity, and cost efficiency in exchange trading. APIs & Integrations Integrate EP3 with third-party or proprietary market access and trading applications. Industry-standard, developer-friendly APIs enable you to integrate custom trading applications, market access connectivity, more efficient order management, clearance and settlement apps, regulatory reporting, and more. Real-time and end-of-day alerts help exchange operators detect unusual activity. EP3’s built-in market surveillance tools reduce complexity by enabling you to trim the number of third-party tools connected to your platform. All Markets & Trading Environments EP3 is built to accommodate all types of markets and trading environments, from new marketplaces to traditional regulated exchanges. - Designated Contract Markets (DCM) - Swap Execution Facilities (SEF) - Request for Quote (RFQ) Markets - Auction Markets - Alternative Trading Systems (ATS) - New Marketplaces - Non-Regulated Market Venues All Asset Classes Our powerful, asset-agnostic technology serves recognized asset classes and a broad range of assets that have never been exchange-traded before. - Futures, Options, Swaps - Environment Commodities (ESG) - Predictions and Events - Equities, Fixed Income, Rates - Physical Commodities - New Asset Classes The MaterialsXchange team had deep domain expertise in the raw materials sector but needed a partner who could build matching engine technology to fit into this ecosystem — and build it quickly to prove the concept. Discover how our EP3 platform enabled MaterialsXchange to launch a robust, reliable exchange in a matter of months. Capitalize On Every Opportunity Earlier Let us show you how our advanced exchange platform and matching engine can propel your organization to the front of the trading race.
systems_science
https://www.charlestonhome.com/articles/ting---electrical-fire-prevention
2024-02-26T11:03:35
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474659.73/warc/CC-MAIN-20240226094435-20240226124435-00710.warc.gz
0.928918
292
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__60565018
en
Ting is an advanced electrical monitor device designed to enhance home fire safety. This innovative device goes beyond traditional fire safety measures by focusing on monitoring electrical systems. Its primary function is to detect and alert homeowners to potential electrical issues that could lead to fires. Equipped with state-of-the-art sensors and intelligent technology, Ting constantly monitors electrical parameters, identifying anomalies or overheating in the electrical system. This proactive approach enables early detection of electrical problems, allowing homeowners to address issues before they escalate into fire hazards. Ting's real-time alerts and notifications provide timely information, empowering residents to take immediate action and prevent potential disasters. The importance of Ting lies in its ability to prevent electrical fires, which are a significant risk in homes. Faulty wiring, overloaded circuits, or other electrical malfunctions cause many fires. Ting is a crucial line of defense, offering a comprehensive solution for monitoring and safeguarding the electrical infrastructure. Installing Ting in your home is a proactive step towards fire prevention. Its user-friendly design and integration into smart home systems make it accessible for homeowners with varying technical expertise. By investing in Ting, homeowners prioritize the safety of their property and loved ones, ensuring a safer living environment by addressing electrical issues before they lead to potential fire incidents. Overall, Ting is a vital device that empowers homeowners to take control of their electrical safety and significantly reduces the risk of electrical fires in the home.
systems_science
https://www.turckbanner.fr/fr/recording-the-swivel-movement-of-a-core-shooter-8147.php
2019-10-23T15:28:19
s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570987834649.58/warc/CC-MAIN-20191023150047-20191023173547-00052.warc.gz
0.84626
162
CC-MAIN-2019-43
webtext-fineweb__CC-MAIN-2019-43__0__188804567
en
Recording the Swivel Movement of a Core Shooter Short cycle times are particularly important when manufacturing sand cores for metal casting using core shooters. The swivel movement of the core carrier, which is recorded by the contactless IO-Link encoder QR24-IOL from Turck, has a major influence on the clock rate of the machine. The rotary encoder detects the overall swivel movement and is therefore able to dynamically control this movement. IO-Link minimizes wiring, facilitates electrical planning, and enables errors to be diagnosed in a transparent manner. - The QR24-IOL contactless IO-Link rotary encoder detects swivel movement and thereby improves the clock rate of machines - Straightforward wiring - Less prone to errors and quicker at detecting errors
systems_science
http://arenabd.com/
2016-10-23T20:16:00
s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988719416.57/warc/CC-MAIN-20161020183839-00337-ip-10-171-6-4.ec2.internal.warc.gz
0.911583
173
CC-MAIN-2016-44
webtext-fineweb__CC-MAIN-2016-44__0__244216211
en
Welcome to Agro Arena Associates Ltd Agro Arena Associates Ltd provides farmers with a wide range of technologies to improve sustainable agriculture for a growing world. Our products and services are designed to solve pressing crop production problems for our farmers; boosting crop productivity to maximum sustainable levels to keep pace with the growing needs of our world’s rapidly expanding population, through higher yields, better varieties, and more targeted pest management control. Our research focus is on game changing technologies to provide better crops, better plant nutrition and better control of destructive crop and noncrop weed and insect pests. The goal of our research efforts is faster, better, more efficient, more productive and more viable agriculture long-term, because our world’s future depends upon crop production technology. Our company enjoys a convenient traffic environment and favorable business surroundings, consequently external resources like information are easily accessible.
systems_science
https://mywishings.com/troubleshooting-and-resolving-voltage-drop-challenges/
2024-04-13T00:06:36
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296816465.91/warc/CC-MAIN-20240412225756-20240413015756-00321.warc.gz
0.858657
896
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__171385172
en
SELECT THE WORDS & LEVEL Voltage drop poses a prevalent challenge in electrical systems, impacting equipment performance and safety. This blog post offers an in-depth exploration of troubleshooting techniques and effective strategies for resolving voltage drop challenges to ensure optimal electrical functionality and longevity. Understanding Voltage Drop and Its Implications Voltage drop occurs when the voltage supplied to an electrical load diminishes as it travels through conductors and connections. Excessive voltage drop can lead to diminished equipment performance, overheating, and compromised operational efficiency, underscoring the necessity for proactive troubleshooting and resolution of voltage drop challenges within electrical installations. Recognizing the Symptoms of Voltage Drop One of the primary steps in addressing voltage drop is recognizing the symptoms associated with it. Flickering lights, electrical equipment that operates slower than usual, or the unexpected shutdown of electronic devices can all indicate a potential voltage drop issue. It’s crucial for technicians to identify these warning signs early to prevent the voltage drop from causing more significant harm to electrical components or leading to unsafe operating conditions. Conducting Comprehensive Load Analysis A foundational step in troubleshooting voltage drop challenges involves conducting a comprehensive load analysis to ascertain the precise demands imposed by electrical equipment and devices. By assessing the current draw, impedance, and voltage requirements of interconnected loads, stakeholders can identify potential sources of excessive voltage drop and devise tailored mitigation strategies to restore optimal electrical performance. Leveraging Advanced Diagnostic Instruments The utilization of advanced diagnostic instruments, including digital multimeters, thermal imaging cameras, and power quality analyzers, empowers technicians to pinpoint areas of high resistance, voltage irregularities, and voltage drop within electrical circuits. Through meticulous diagnostic assessments, operators can discern anomalous voltage behaviors, localize problematic components, and initiate targeted rectification measures to alleviate voltage drop complications. Implementing Corrective Wiring and Connection Practices Corrective wiring and connection practices play a pivotal role in ameliorating voltage drop challenges and enhancing electrical system stability. From optimizing conductor sizing to minimizing the length of circuitous pathways and fortifying connections, adhering to best practices in wiring installation mitigates resistive losses, attenuates voltage drop, and bolsters the integrity of electrical pathways, thereby fostering sustained operational efficacy. Integrating Submersible Pressure Transducers for Accurate Measurements In addressing voltage drop issues within systems exposed to fluids or submerged environments, integrating submersible pressure transducers can be instrumental. These robust sensors are expertly crafted to withstand harsh, wet conditions and provide accurate measurements in real-time, which is essential for monitoring the effects of voltage drop on equipment operating underwater or fluid. By deploying submersible pressure transducers, technicians can obtain precise diagnostic insights, enabling them to make informed decisions when adjusting or repairing electrical systems to counter the nuances of voltage drop in such unique scenarios. Integrating Voltage Stabilization Technologies The integration of voltage stabilization technologies, such as automatic voltage regulators (AVRs) and power conditioners, furnishes an effective mechanism for mitigating voltage drop and sustaining consistent voltage levels across electrical networks. By deploying voltage regulation apparatus at pertinent junctures, organizations can safeguard against voltage fluctuations, temper voltage drop manifestations, and fortify the resilience of sensitive electrical equipment to erratic voltage variations. Enhancing Conductor Material Selection Conductor material selection plays a pivotal role in mitigating voltage drop challenges, particularly in applications characterized by extended cable runs and high current loads. Opting for conductors with reduced resistance properties, such as copper or aluminum alloys with enhanced conductivity, curtails resistive losses, minimizes voltage drop, and optimizes the transmission of electrical power within diverse operational contexts. Proactively Addressing Environmental Factors Proactively addressing environmental factors that exacerbate voltage drop is essential for preserving the stability and longevity of electrical systems. By shielding cables from extreme temperature fluctuations, mitigating the impact of ambient humidity, and safeguarding against corrosive elements, organizations can preemptively attenuate the influence of environmental variables on voltage drop, thereby fostering sustained electrical reliability. Troubleshooting and resolving voltage drop challenges demand a comprehensive approach encompassing load analysis, diagnostic instrumentation, corrective wiring practices, integration of voltage stabilization technologies, judicious conductor material selection, and proactive environmental management. By proactively addressing voltage irregularities and implementing tailored mitigation strategies, organizations bolster the operational resilience of electrical systems, optimize equipment performance, and uphold the safety and functionality of electrical installations across diverse industrial domains.
systems_science
https://kelseymartinlab.com/research/synapse-specific-memories
2023-09-23T12:12:06
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233506480.7/warc/CC-MAIN-20230923094750-20230923124750-00401.warc.gz
0.908694
211
CC-MAIN-2023-40
webtext-fineweb__CC-MAIN-2023-40__0__286809676
en
Since each neuron has a single nucleus but can form thousands of synaptic connections, the requirement for transcription during synaptic plasticity raises the question of how the products of gene expression can be targeted to alter synaptic strength at select synapses made by a given neuron. We have found that one important mechanism involves the translation of synaptically localized mRNAs. Current research addresses the following questions: 1) what is the population of localized mRNAs in neurons? 2) how do mRNAs localize to synapses—what are the cis-acting RNA localization elements and the trans-acting RNA binding proteins? 3) how does synapse formation and/or synaptic stimulation regulate either the trafficking or the translation of localized transcripts? 4) what is the function of local translation in memory formation? From a broad perspective, we would like to understand how gene expression can be spatially and temporally regulated at the level of the synapse, and how this local translational regulation is coordinated and integrated with transcriptional changes in the neuron.
systems_science
https://blog.massdrive.com/2010/03/05/should-we-automate-toll-booths/
2024-04-20T13:22:21
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817650.14/warc/CC-MAIN-20240420122043-20240420152043-00018.warc.gz
0.97747
252
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__142661894
en
Highway officials may start testing a new form of electronic tolling in the next couple months. Governor Patrick Deval announced in his monthly radio station chat test runs of the proposed system may appear at the Massachusetts Turnpike at the Route 128 exchange. With the new system drivers would no longer have to pull up to toll booths, rather they would drive under a scanner that would read transponders like those currently utilized by FastLane drivers. However the electronic tolling would also have the ability to read a license plate and send a bill or charge a driver’s account. Governor Patrick Deval believes the new system will make toll-paying easier and safer for drivers as they would no longer need to fumble for change. Although the FastLane transponders are given out free and payment lanes are available, some drivers still prefer to pull up to a booth. A cash lane would need to be kept for those who do not have transponders, such as out of state drivers or new vehicles. Deval said this past Wednesday he would like to have the new system trials up and running sooner then later. When asked to elaborate on a time period Deval stated he hoped to see testing in the next few months.
systems_science
http://www.manifold.net/doc/mfd9/microsoft_office_formats_-_mdb,_xls_and_friends.htm
2020-07-14T02:29:43
s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593657147917.99/warc/CC-MAIN-20200714020904-20200714050904-00279.warc.gz
0.871846
3,773
CC-MAIN-2020-29
webtext-fineweb__CC-MAIN-2020-29__0__216562234
en
When our Windows system has the required Microsoft components installed, reading and writing Microsoft Office formats like Access MDB or Excel XLS will be effortless in Manifold. The trick is sorting out the Microsoft installation. That can be difficult because of conflicts between different Microsoft packages that have evolved over time in incompatible ways. If we sort out our Microsoft installation, everything works perfectly, all of the time. Important: This topic applies to all Office formats managed by Microsoft facilities, including legacy Office formats such as .db, .html. .mdb, .xls, and .wkx, together with newer Office formats such as .xlsx and .accdb. This topic also applies to any other use of .mdb format, for example, ESRI's personal geodatabase format using .mdb or legacy Manifold mfd/mdb format as used in Manifold Release 4.50. Manifold Viewer is a read-only version of Manifold, so everything in this topic applies not just to Manifold System but also to the free Viewer product from Manifold. Manifold utilizes Microsoft data access software to connect to Microsoft Office file formats. Using Microsoft's own code ensures that data will be read from and written to those formats as Microsoft intends, and it automatically takes advantage of Microsoft bug fixes and improvements in Microsoft updates. However, when using Microsoft's code we are subject to Microsoft's requirements for compatibility between different Microsoft products. For connections to Office formats, Manifold can use any of the three data access systems Microsoft has used over the years to support Microsoft Office. Only the last of the three, Access Database Engine, is recommended: JET - JET is a 32-bit package that was Microsoft's original data access technology for Office, but which is now deprecated in favor of newer technologies. JET can read legacy Microsoft Office files in formats such as .db, .html. .mdb, .xls, and .wkx if they have been created by JET, but might not be able to read them if they were created using newer Microsoft technology like ACE or Access Database Engine. JET cannot read or write newer Office formats such as .xlsx and .accdb. Because JET is 32-bit, to use it we must launch Manifold in 32-bit mode. This applies even if we are using a 64-bit Windows system. JET is not recommended. Instead, we should use Access Database Engine. Office Access Connectivity Engine (ACE) - Introduced in 32-bit form in 2007, ACE introduced revised versions of old formats such as .mdb and also introduced new Office formats such as .xlsx and .accdb. A 64-bit version of ACE was introduced in 2010 to support 64-bit Office. ACE is basically an old version of Access Database Engine that should be replaced with the latest version. Access Database Engine - The new name for ACE, as of the present writing the Access Database Engine is available in either a 2010 version or a 2016 version, both versions being available both as 32-bit or as 64-bit products. Access Database Engine is the best way to connect to Microsoft Office formats, when it works. 64-bit Access Database Engine allows 64-bit Manifold to connect to Office formats, including legacy formats like .mdb and newer formats like .xlsx and .accdb. Because of severe flaws in Access Database Engine installation and compatibility with other Microsoft software, Manifold does not automatically install Access Database Engine. We must download it from Microsoft's web site and install it. Access Database Engine is a free download. Our objective: In an ideal world, when running 64-bit Manifold on a 64-bit Windows system we should install the latest version of Microsoft's 64-bit Access Database Engine. That will allow 64-bit Manifold to connect to Office formats. When running 32-bit Manifold on a 32-bit Windows system, we should install the latest version of Microsoft's 32-bit Access Database Engine. That will allow 32-bit Manifold to connect to Office formats. However, due to flaws in Access Database Engine installation and compatibility with other Microsoft applications, we may be unable to install or use Access Database Engine. If we cannot get Access Database Engine to work, or if we find the procedures to deal with Access Database Engine installation flaws and operational bugs too inconvenient, we can launch and use a 32-bit instance of Manifold. That always works for the formats JET supports, but does not provide connectivity to newer formats like .xlsx or .accdb. There is no sugar-coating it: installation of 64-bit Access Database Engine can be an inconvenient mess. It is not compatible with 32-bit Microsoft Office, and it depends upon Microsoft libraries which might not be on the Windows system and which might not be installed by the Access Database Engine installation. In such cases the installation will not fail in a clean way, but instead operation of the Access Database Engine will be faulty, with failed operations. Other bugs in Access Database Engine may cause crashes in matters entirely unrelated to reading and writing Microsoft formats like .mdb and .xls. It really is a mess, but if we are persistent and go step by step we can usually get it running. A Microsoft link to download the Microsoft Access Database Engine 2016 Redistributable package is: https://www.microsoft.com/en-us/download/details.aspx?id=54920. Microsoft links may change. If so, use your Internet search engine to search for "Microsoft Access Database Engine 2016 Redistributable" to find a current download link. The above download provides downloads for an x64 (64-bit) Access Database Engine as well as a 32-bit version. Most people only install the x64 version since 64-bit Manifold is where it is used. In 32-bit Windows systems our task is simple: download and install 32-bit Access Database Engine. In 64-bit Windows system we may encounter a complication: 64-bit Access Database Engine cannot be installed on a 64-bit Windows system if 32-bit Microsoft Office has been installed. That can be a problem because many people use 32-bit Microsoft Office installations within 64-bit Windows. If we try to install 64-bit Access Database Engine on such a system, we get an error message: We can fix the above by downloading and installing 32-bit Access Database Engine. We can then launch Manifold in 32-bit mode whenever we want to import from or export to .mdb or other Office formats, like .xls, .xlsx, or .accdb. We can import an .mdb in a 32-bit Manifold session and save the .map project. We can then open that .map project in a 64-bit Manifold session to get the benefits of 64-bit power. If we must leave data in a linked .mdb file instead of importing it into Manifold, it is no problem to use 32-bit Manifold, because .mdb files are limited to only 2 GB, a small size that can be easily handled by 32-bit Manifold. In both 64-bit and 32-bit versions, Access Database Engine depends upon a Microsoft library called MSVCR100.DLL. However, as of this writing (2020) the Access Database Engine installation package fails to install that library if it is not present on our computer. If the library is not present, Access Database Engine will not warn us. Instead, it will try to load another DLL that depends on the missing library, which fill fail. The operation that called that DLL, which is usually a reading operation, will abort. In export cases, it will keep trying to load the DLL that depends upon the missing MSVCR100.DLL and that will fail repeatedly. That may abort the process or the export might work but only very slowly. To solve that problem we can search the Microsoft site for the missing DLL and install it. Search for both the 64-bit and 32-bit versions of the DLL and install both. The current names of the missing DLL in 64-bit and 32-bit versions are: Microsoft's site changes frequently so reliable links to the above cannot be published in this documentation. Instead, use the search feature of the Microsoft site to find them, or use a search engine such as Bing or Google to find them. Crazy as it sounds, the Access Database Engine uses graphics services, and if hardware-acceleration for graphics is turned on within the Access Database Engine it can crash in some versions. This bug has been acknowledged by Microsoft and has been said to be fixed on past occasions but without being fixed in all circumstances. A Microsoft link that discusses the issue: https://docs.microsoft.com/en-us/office/troubleshoot/access/odbc-administrator-crashes Although the bug may be fixed in whatever is the current version of the Access Database Engine, the safe plan is to turn off hardware graphics acceleration within Access Database Engine dialogs, just in case. That must be done by editing the Windows registry. Edit the registry to alter this key: ... to this value: (DWORD) DisableHardwareAcceleration = 1 Editing the Windows registry can cause system problems if done incompetently. We should not edit the registry if we do not know how to do so competently. In a 64-bit Windows system it would be more convenient to simply launch 64-bit Manifold to connect. But if that is not possible due to the presence of 32-bit Office, we can live with having to use 32-bit Manifold when importing from or exporting to .mdb, .accdb, .xls, .xlsx, and other Office formats. Yes, that is a hassle, but it is less of a hassle than a choice between abandoning Office or buying 64-bit Office. Since a 64-bit Manifold license includes both 64-bit and 32-bit Manifold, we can use 32-bit Manifold to get around the conflict between Microsoft's 32-bit Office and Microsoft's 64-bit Access Database Engine. To launch Manifold in 32-bit mode in a 64-bit Manifold system, see the Launch in 32-bit Mode topic. Access Database Engine already may be installed on our Windows system. The easiest way to find out is to launch Manifold and then to try linking an .xlsx or .accdb file. We try linking an .xlsx or .accdb file, because if we try linking an .mdb the result does not tell us if we have JET or if we have Access Database Engine. If we have a 64-bit Manifold installation, we try linking an .xlsx or .accdb file twice, first launching Manifold in 64-bit mode to try linking, and then again launching Manifold in 32-bit mode to try linking. That will tell us if we have 64-bit or 32-bit Access Database Engine installed. If we are curious to see if we have JET installed, we can launch Manifold in 32-bit mode and also try linking an .mdb file, to see what happens. 32-bit Windows - If we have 32-bit Windows, Manifold is always 32-bit, just like everything else on our system. We launch Manifold and use File - Link to try linking an .xlsx or .accdb file. If we can link an .xlsx or a .accdb file, that means we have Access Database Engine installed. Done! If we cannot link .accdb or .xlsx files, that means we do not have Access Database Engine installed. Download 32-bit Access Database Engine from the Microsoft web site and install it. 32-bit Windows, running JET - If we can link an .mdb within Manifold but not an .accdb or .xlsx, that means we have JET installed but not Access Database Engine. JET is so ubiquitous that most older Windows systems, or older Windows systems that were automatically upgraded to Windows 10, have JET installed. If JET is installed, Manifold will happily use it to import legacy .mdb and other legacy Office files that JET can handle. Many Manifold users are running on JET without realizing it: they launch 32-bit Manifold to import an .mdb and it works, so no need to get into why it works. However, it is unwise to run on very old, deprecated Microsoft data access software like JET. The right move is to forget about JET and to download and to install Access Database Engine. 64-bit Windows - Launch Manifold in 64-bit mode and use File - Link to try linking an .xlsx or .accdb file. If we can link an .xlsx or a .accdb file, that means we have 64-bit Access Database Engine installed. Done! If that does not work, launch Manifold in 32-bit mode and again use File - Link to try linking an .xlsx or .accdb file. If that works, we have 32-bit Access Database Engine installed. If 32-bit Access Database Engine is installed, we should use it, launching Manifold in 32-bit mode whenever we want to connect to Office format files. If neither 64-bit nor 32-bit Manifold can link an .xlsx or an .accdb file, we do not have Access Database Engine installed. Download the 64-bit Access Database Engine from the Microsoft web site and install it. If an error message appears saying 64-bit Access Database Engine cannot be installed due to the presence of 32-bit software, download and install the 32-bit Access Database Engine, and then use Manifold in 32-bit mode to connect to Office files. It must be emphasized that the above integration issues arise from conflicts various Microsoft packages have with other Microsoft packages. Manifold is happy to use whatever Microsoft facilities we have installed. Early Manifold releases, such as Manifold System Release 4.50, utilized Microsoft .mdb format as an integral part of Manifold's vector storage format. Manifold mfd/mdb format utilizes two files, a Manifold .mfd file to store geometry and a Microsoft .mdb file to store attributes. Early Manifold releases that utilized mfd/mdb format automatically installed a copy of Microsoft JET to provide data access to the .mdb file. Modern Manifold releases do not install JET, and depend on the installation of Microsoft data access facilities as discussed in this topic. If we do not have .mdb capability installed in our Windows system as discussed in this topic, for example, by installing Access Database Engine, Manifold System will not be able to import the .mdb portion of legacy Manifold mfd/mdb format. Drawings in mfd/mdb format will be imported with only object geometry, but without attributes, that is, without the data fields for each object. Can't install 32-bit Access Database Engine - In 64-bit Windows systems that have 32-bit Office installed, Microsoft will not allow installation of 64-bit Access Database Engine. The usual workaround is to install 32-bit Access Database Engine. However, that sometimes fails, with mysterious error messages about "64-bit Office products" being installed. A diligent Internet search will usually produce a solution. For an example, see this thread on Microsoft's site. Does Manifold include JET? - No. Early versions of Manifold, such as Release 4.50, included JET to support Manifold's own mfd/mdb format. Manifold later dropped JET and shifted to Manifold's own 64-bit DBMS format, relying on Microsoft installations of JET to support data access to .mdb and other Office formats. Why doesn't Manifold automatically install Access Database Engine? - Three reasons: First, Access Database Engine would increase the size of the Manifold installation by over 160 MB, which would be unfair to people who do not intend to connect to Office formats or who already have Access Database Engine installed. Second, integration issues between Access Database Engine and other Microsoft software are potentially too complicated to be handled by a Manifold installation script. If Access Database Engine is not already installed in a Windows system, it is a very simple matter to get it from the Microsoft web site and install it, for those users who want to connect to Office formats. Finally, despite Manifold's advice to switch from JET to Access Database Engine, many people already have JET installed on their systems and are happy using JET for Manifold connections to .mdb and similar. Access connecting to Manifold - When linking Manifold tables into Access using the Manifold ODBC driver, please review the notes on Access in the DBMS Data Sources - Notes topic. Microsoft is not Manifold - As convenient as it is to be able to use Microsoft's own code to connect to Microsoft's products, when using facilities like Access Database Engine we should keep in mind we are running MIcrosoft code. Although Microsoft code has a good reputation, in general it is not as bulletproof as the Radian technology used within Manifold. Connections using Access Database Engine fall outside of Manifold's reputation for never crashing. MDB Microsoft Access Example: Create and Use New Data Source using an MDB Database - This example Illustrates the step-by-step creation of a new data source using an .mdb file database, followed by use of SQL. Although now deprecated in favor of the more current Access Database Engine formats, .mdb files are ubiquitous in the Microsoft world, one of the more popular file formats in which file databases are encountered. Launch in 32-bit Mode
systems_science
https://www.backbone.vc/post/bbv-miniseries-2-megatrend-ai-automation-3-use-cases
2024-03-04T17:10:51
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947476464.74/warc/CC-MAIN-20240304165127-20240304195127-00066.warc.gz
0.933915
839
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__160407888
en
Artificial Intelligence will be a key technology of the 21st century. We take a look at our portfolio companies and explain trends and use cases that will reshape our economy and society in the decades to come. When James Hargreaves invented the Spinning Jenny in 1764, little did he know that his invention would change the world forever. Mechanized spinning was one of the core innovations that triggered the first wave of industrialization, which resulted in a significant increase in automation and manufacturing output. Today, 256 years later, we are in the early stages of the fourth industrial revolution — or “Industry 4.0” — which again will increase the level of automation. But what technologies will dominate, and which company will invent the Spinning Jenny of the 21st century? In part #2 of our miniseries on Megatrends, we will take a look at Artificial Intelligence (AI). Autonomous transportation with AI-powered drones Transportation will look entirely different at the end of this century compared to the status quo. Already today, big tech is working on autonomous cars and drones that can deliver goods, take pictures, and transport human passengers. The use of AI in transportation will gradually decrease the need for operator intervention, and supervision and the complexity of actions that vehicles can perform will significantly increase. Our portfolio company Flytrex runs a drone-based delivery service in the suburbs of North Dakota and Iceland. Drones can carry up to 6.6 pounds (6–8 hamburgers) for a distance of 3.5 miles and back. The company’s success shows that there is demand for such a solution, as customers value convenience and low costs. Drones can fly on autopilot on a programmed route, called waypoints, and perform certain actions. In the future, AI will enable drones to become more capable of performing more complex operations at scale. Instead of delivering hamburgers in a box, drones will then carry out international cargo flights and eventually also passenger flights. Watch the sky; drones will soon become the new normal. Robotic Process Automation (RPA) RPA combines software robots (bots) and AI to automate workflows that were previously performed by human beings. In contrast to traditional workflow automation, where a developer programs software to perform certain tasks, RPA-bots observe a human while performing tasks and then simply repeat those exact same tasks. Simply put: Instead of programming a bot, humans will train and teach it, enabling it to learn and expand its capabilities autonomously. Ui Path, for example, uses bots to automate all kinds of business processes, from accounting to customer services and human resources. Their bots can take over every process that follows a repetitive routine — for example, insurance claims processing — by observing how humans perform the process steps. Once the bot has learned each process step, it will be able to take over and perform the entire process autonomously. AI-based financial decision-making While most automation focuses on routine processes, AI also plays an increasingly important role in automated decision-making. Wherever decisions are made based on data, AI can streamline the data ingestion and processing to build models without human intervention. Take finance, for example. Many decisions regarding investments or loan-processing are purely based on financial data. Zest AI has developed an AI-based system that autonomously analyses credit scores, past payment behaviour, and cashflows to enable banks and other lenders to make a fast but precise lending decision. Instead of a bank employee manually analysing credit data, AI takes over the entire credit analysis, and a human only makes the final lending decision. So, will AI be the new Spinning Jenny? Possibly. Elon Musk believes AI will become so powerful that he even calls it a “fundamental risk to the existence of human civilization.” But we are not there yet. Before we arrive at a Matrix-like doomsday scenario, AI will provide us with many benefits. The start-ups above are just some examples of the gems that are already working on our automated future.
systems_science
https://lilvtechs.com/blockchain/
2023-05-30T00:46:09
s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224644915.48/warc/CC-MAIN-20230530000715-20230530030715-00677.warc.gz
0.94728
206
CC-MAIN-2023-23
webtext-fineweb__CC-MAIN-2023-23__0__108870632
en
Blockchain technologies are predominantly related to cryptocurrencies such as Bitcoin, but there’s a lot more to the technology than digital currency. The security it offers is useful in lots of ways. Perhaps the best way to describe Blockchain is that data which can only be added to, not taken away from or changed. This is where the suffix ‘chain’ comes from, as what is being created in the process is a ‘chain’ of data. The key highlight about blockchain is that previous blocks of data already created cannot be changed, making it a very secure option. Also, blockchains are consensus driven, meaning no one entity gets control of the entire data. In Blockchain, there is no need for a third party to oversee or validate transactions. Studies reveal that Blockchain related jobs are some of the fastest growing ones, with 14 openings for every one blockchain developer in the offing. Blockchain developers use blockchain technology to develop and implement architecture and solutions, their average median salary being $130,000.
systems_science
https://preishragency.com.ua/jobs/process-and-systems-analyst/
2022-06-28T06:44:46
s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103355949.26/warc/CC-MAIN-20220628050721-20220628080721-00271.warc.gz
0.904967
858
CC-MAIN-2022-27
webtext-fineweb__CC-MAIN-2022-27__0__262710402
en
The Process and Systems Analyst will be responsible for the design and implementation of processes and systems improvement within the company’s SDLCs in tight collaboration with other company organizations, functions, and departments. The Process and Systems Analyst will be working closely with the company’s leaders on reaching strategic objectives of PMO, Portfolio functions, Solutions organization, and the whole company. The Process and Systems Analyst will be directly reporting to the Process and Systems Analyst Lead as part of the Process Development Team unit, PMO. In this role you will get to: Primarily responsible for company SDLCs’ maintenance; Owns a backlog of company SDLCs’ change requests, performs change requests evaluation, impact analysis, and further action planning; Participates in design and implementation of processes and systems improvements within company’s SDLCs, partnering with process owners from PMO and other departments from Portfolio function, Solutions organization and globally; Develops process, procedures, policies, and systems improvements within the company’s SDLCs; Proactively interacts with business process owners to collect feedback and plan processes and systems improvements; Communicates change in processes and systems to key stakeholders and provides needed training and support; Actively participates in internal PMO development initiatives, shares knowledge, and experience to reach initiative’s objectives; Proactively identifies and efficiently resolves issues in consultancy with other departments if needed; Participates in risks identification and analysis of processes and systems improvements within the company’s SDLCs; Establishes and efficiently manages communications with business owners and other key stakeholders in a frame of processes and systems improvements within the company’s SDLCs; Proactively communicate status, issues, and risks, related to improvements of the company SDLCs’ processes and systems, to business owners and key stakeholders; Owns decision-making within a defined level of authority. Proactively and efficiently escalates out of authority issues and risks to upper-level management if needed. Your skills and qualifications: Bachelor’s degree or higher in Business Administration or Organizational Development; or equivalent work experience is a plus; At least 3 years of experience working with various business applications and systems. Experience with JIRA, Confluence, SalesForce, and PSA is a plus; Good understanding and practical experience with the development of business processes, procedures, policies, and/or large-scale organizational change efforts in mid-large size organizations; Ability to speak technical and business languages, and interpret and document requirements; Excellent business analytical and technical writing skills; Excellent oral communication skills, active listening skills; Proven ability to conduct stakeholder interviews in both individual and group settings without additional supervision; Technically competent and able to quickly learn new technologies, self-motivated; Ability to establish and maintain strong relationships with cross-department leaders; Flexibility and adaptability; Good problem-solving skills; Basic understanding of project management areas following the PMBOK; Fluent written and spoken English. What we offer in return: The knowledge base of the world’s largest e-commerce delivery team; Professional training and certifications; Off-the-Charts Career Growth: Clear career path and a performance review system, career coaching, training and certifications, mentoring, and knowledge sharing. Well-being Is Top Priority: Parental leave, paid time off, comprehensive health and medical plans. Real Work-Life Balance: Remote, in-office, or hybrid working modes; flexible hours; work-life balance support on every stage and level. Culture of Success: Culture of collaboration that encourages innovation every step of the way; 20 offices spanning four continents bring diverse perspectives that drive tangible results for our clients worldwide. About the Company: Our client is a global e-commerce agency headquartered in the US with over 1200 professionals in 12 countries.
systems_science
https://duniacm.com/erp-software-for-small-manufacturing-business/
2024-02-24T20:40:00
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474544.15/warc/CC-MAIN-20240224180245-20240224210245-00583.warc.gz
0.92366
1,729
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__90583400
en
For small manufacturing businesses, managing complex operations can be a challenging task. From inventory management to order tracking, there are numerous aspects to consider to ensure that the business is running smoothly. Enterprise Resource Planning (ERP) software can help small manufacturing businesses streamline their operations and improve overall efficiency. ERP software is a business management software that integrates all aspects of a company’s operations, including manufacturing, sales, inventory, and finance. This integration of all business processes into a single system can help companies manage their operations more effectively and efficiently. In this article, we will discuss the benefits of using ERP software for small manufacturing businesses. Benefits of ERP Software for Small Manufacturing Businesses Streamlining Business Processes ERP software can streamline business processes by integrating all aspects of a company’s operations into a single system. This integration can reduce the need for manual data entry and improve the accuracy of business data. For example, when a customer places an order, the order is automatically entered into the ERP system. The system then generates a production order, which is sent to the manufacturing department for production. Once the order is produced, the system automatically updates the inventory and generates an invoice. This automation of business processes can save time and reduce the chance of errors. Improving Inventory Management ERP software can help small manufacturing businesses improve inventory management by providing real-time visibility into inventory levels. This visibility can help businesses manage their inventory more effectively by ensuring that they have the right inventory levels at the right time. Additionally, ERP software can help businesses reduce inventory costs by optimizing inventory levels and reducing excess inventory. Enhancing Customer Relationship Management ERP software can enhance customer relationship management by providing businesses with a centralized database of customer information. This information can include contact details, purchase history, and other relevant information. This centralized database can help businesses provide better customer service by allowing them to quickly access customer information and respond to customer inquiries more effectively. Improving Financial Management ERP software can help small manufacturing businesses improve financial management by providing real-time visibility into financial data. This visibility can help businesses manage their cash flow more effectively by providing information on accounts payable, accounts receivable, and cash balances. Additionally, ERP software can help businesses improve their financial reporting by providing accurate and timely financial reports. Choosing the Right ERP Software for Small Manufacturing Businesses Understanding Business Needs When choosing an ERP software for small manufacturing businesses, it is essential to understand the specific needs of the business. Businesses should consider their business processes, size, and budget when choosing an ERP software. Evaluating ERP Vendors Once businesses have identified their business needs, they should evaluate ERP vendors based on their experience in the manufacturing industry, the functionality of their software, and the level of support they provide. Considering Cloud-Based ERP Software Cloud-based ERP software can provide small manufacturing businesses with numerous benefits, including lower upfront costs, easier implementation, and automatic updates. However, businesses should also consider the security and data privacy risks associated with cloud-based ERP software. Implementing ERP Software for Small Manufacturing Businesses Planning for Implementation Implementing an ERP system can be a complex process, and it is essential to have a plan in place to ensure a successful implementation. The first step is to define the scope of the implementation, including which business processes the system will cover, which departments will be involved, and what the timeline will be. Data migration is another critical step in implementing an ERP system. This involves transferring data from the old system to the new system, ensuring that the data is accurate and complete. It is important to have a data migration plan in place to ensure that the process is efficient and effective. Training is crucial for ensuring that employees can use the new system effectively. This includes providing training on how to use the software and how to perform specific business processes. It is important to provide ongoing training to ensure that employees can continue to use the system effectively and efficiently. Testing and Support Before going live, it is important to test the system thoroughly to ensure that it is working correctly. This includes testing all business processes and ensuring that the system is integrated with other systems, such as accounting software. Once the system is live, it is important to provide ongoing support to ensure that any issues are resolved quickly. Implementing an ERP system can be a significant undertaking for small manufacturing businesses, but it can also provide significant benefits. By streamlining business processes, improving inventory management, enhancing customer relationship management, and improving financial management, ERP software can help small manufacturing businesses improve efficiency, reduce costs, and grow their business. When choosing an ERP software, it is important to consider business needs, evaluate vendors, and consider the benefits and risks of cloud-based software. Planning for implementation, data migration, training, and ongoing support are essential for ensuring a successful implementation. Common Features of ERP Software for Small Manufacturing Businesses Production Planning and Scheduling Production planning and scheduling are critical aspects of small manufacturing businesses. ERP software can help with production planning by providing tools to help businesses plan production based on demand and available resources. This can help businesses optimize their production schedule to ensure that they can meet customer demand and reduce lead times. Inventory management is another important feature of ERP software for small manufacturing businesses. ERP software can help businesses track inventory levels in real-time and provide tools to optimize inventory levels based on demand and production schedules. This can help businesses reduce excess inventory and improve cash flow. Supply Chain Management Supply chain management is another area where ERP software can help small manufacturing businesses. ERP software can provide tools to manage suppliers, track shipments, and monitor delivery times. This can help businesses optimize their supply chain to reduce lead times, improve delivery times, and reduce costs. Quality control is critical for small manufacturing businesses. ERP software can provide tools to track product quality throughout the production process and ensure that products meet customer specifications. This can help businesses reduce product defects, improve customer satisfaction, and reduce costs associated with product recalls. Financial management is another essential feature of ERP software for small manufacturing businesses. ERP software can provide tools to manage accounts payable, accounts receivable, and cash flow. Additionally, ERP software can provide real-time visibility into financial data to help businesses make informed financial decisions. Advantages of Implementing ERP Software for Small Manufacturing Businesses Improved Efficiency and Productivity One of the most significant advantages of implementing ERP software for small manufacturing businesses is improved efficiency and productivity. ERP software can automate business processes and reduce manual data entry, freeing up employees’ time to focus on higher-value activities. This can lead to improved productivity and increased output. Enhanced Customer Relationship Management ERP software can also help small manufacturing businesses improve customer relationship management. By providing real-time visibility into customer orders and delivery schedules, businesses can improve customer service and respond more quickly to customer inquiries. This can lead to increased customer satisfaction and repeat business. Better Inventory Management Inventory management is another area where small manufacturing businesses can benefit from implementing ERP software. By providing real-time visibility into inventory levels and optimizing inventory based on demand and production schedules, businesses can reduce excess inventory, improve cash flow, and reduce costs associated with carrying inventory. Improved Financial Management ERP software can also help small manufacturing businesses improve financial management. By providing real-time visibility into financial data, businesses can make informed financial decisions and improve cash flow. Additionally, ERP software can automate financial processes such as invoicing and accounts payable, reducing the time and effort required to manage these processes manually. Better Decision Making Finally, implementing ERP software can help small manufacturing businesses make better decisions. By providing real-time visibility into key business data, businesses can make informed decisions based on accurate and up-to-date information. This can lead to improved profitability and growth. Despite the potential challenges associated with implementing ERP software, the benefits for small manufacturing businesses can be significant. Improved efficiency and productivity, enhanced customer relationship management, better inventory management, improved financial management, and better decision-making are just a few of the potential benefits. When considering implementing ERP software, small manufacturing businesses should carefully evaluate the potential challenges and ensure that they are prepared to invest the time and resources necessary for a successful implementation.
systems_science
https://stasislabs.com/rpm-solution/
2023-12-09T02:53:40
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100781.60/warc/CC-MAIN-20231209004202-20231209034202-00183.warc.gz
0.909737
649
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__107133803
en
We believe that patient-clinician connection shouldn’t only happen next to the bedside. With Stasis, clinical staff gains access to remotely see how their patients are doing with the assurance that they’re connected when they need to be. With the Stasis Connectivity Platform, you can see vital signs monitor data for all patients, seamlessly broadcasted wherever you need it: nursing stations, observation areas, or on-the-go from a smartphone. High acuity areas, such as Emergency Departments, Step-Down Units, and specialty areas, see an overwhelming load of different patients and conditions each day. With Stasis, you can set up observation dashboards in central areas to track all your patients remotely and use automatic risk sorting to highlight high-risk cases. Customizable notifications & alerts can be personalized for each patient, reducing alarm fatigue while maintaining safety. Critical events or sudden deterioration can happen in low acuity areas, such as in private rooms and wards, where clinical surveillance is the lowest. With the Stasis web and mobile apps, care teams can keep track of their patients’ statuses from wherever they are, preventing delays in treatment or missed events. Insight reports help to visualize the complete patient profile throughout their stay and assist with charting from admission to discharge. Outpatient care sees a high volume of patients coming through the doors. Stasis helps ease the burden of manual admin tasks for your healthcare staff with automated data collection from vital signs. Integrations with HIS/EMR systems allow clinicians to access vital sign information without having to switch between multiple systems. In an outpatient setting, where patients come are often seen by multiple clinical staff members, data integrations guarantee that all patient information is accurate, complete, and up-to-date, allowing for better coordination of care. The Stasis Platform provides paramedics with a powerful tool to monitor, track and communicate patient information in real time. By automating and digitizing data collection, paramedics can spend more time focused on patient care, rather than on documenting information. By providing hospital staff with remote access to vitals information, they can be better prepared to receive the patient when they arrive, leading to a smoother transition of care. * As seen in an independent 5-month peer-reviewed published clinical study using Stasis in a hospital step-down unit. You plug us in — we provide you with a smart connectivity solution that makes it easy to extract clinical and operational value from your data. Here’s how: Automate data collection from vital signs monitors. By digitalizing your charting process, you can eliminate manual charting, decrease errors, and save time, while enhancing clinical decision-making with immediate access to your data. Send and receive data with HIS/EMR using a flexible cloud infrastructure that meets your specific privacy and security requirements. Key patient data is now available for review, digitally, in the platform that is most convenient for you: within the hospital information system, other hospital software, and on the Stasis secure RPM apps. Listen to what the doctors in India have to say about how the Stasis Platform has facilitated seamless patient care in high-risk wards.
systems_science
https://en.joinlets.de/datenschutz
2024-04-18T00:03:26
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817184.35/warc/CC-MAIN-20240417235906-20240418025906-00684.warc.gz
0.955983
259
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__134474841
en
lets is based on state-of-the-art technology to ensure maximum protection for you and your data. For this reason, all data transmissions are SSL-encrypted and data is only stored on German servers. The data is encrypted with Transport Level Security (TSL) during transmission between us and your browser. All data is stored exclusively on German servers. This means we are subject to strict data protection regulations - and we think that's a good thing! lets is DSGVO compliant and built to facilitate privacy protection through simple design and clear privacy policies and features. This way, you as a user are always in control of your data. Our data centres are ISO-certified and therefore independently rated as secure. We always try to store as little data as possible so that a great user experience is still possible, but no data is stored without meaning. You can decide yourself which data you want to publish and which data you prefer to share personally. Decide whether you want to use our two-factor authentication when logging in. A second code is sent and requested via SMS when you log in. To ensure the security of your data, we subject our systems to regular code scans and pentests. This ensures that we don't miss a single detail.
systems_science
https://asabe.figshare.com/articles/figure/Supplemental_figures_for_Mass-Balance_Process_Model_of_a_Decoupled_Aquaponics_System_/22788446/1
2023-10-04T12:19:14
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233511369.62/warc/CC-MAIN-20231004120203-20231004150203-00030.warc.gz
0.701244
171
CC-MAIN-2023-40
webtext-fineweb__CC-MAIN-2023-40__0__264154287
en
Supplemental Material.pdf (119.63 kB) Supplemental figures for "Mass-Balance Process Model of a Decoupled Aquaponics System" figureposted on 2023-08-11, 17:50 authored by Rohit Kalvakaalva, Mollie Smith, Stephen A. Prior, G. Brett Runion, Emmanuel Ayipio, Caroline Blanchard, Daniel Wells, David M. Blersch, Sushil Adhikari, Rishi PrasadRishi Prasad, Terrill R. Hanson, Nathan R. Wall, Brendan Higgins Figure S1 shows water and feed inputs to the aquaponics system over the course of one year. Figure S2 shows modeled rates of nitrogen gas evolution over time due to denitrification in the aquaponics system.
systems_science
https://www.virtualdata.com/remote-desktop-services/
2021-12-02T14:54:28
s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964362230.18/warc/CC-MAIN-20211202145130-20211202175130-00135.warc.gz
0.9122
177
CC-MAIN-2021-49
webtext-fineweb__CC-MAIN-2021-49__0__20919112
en
Have your Desktop and Applications accessible from anywhere, on any device and at any time. Simplicity for setting up a secure remote workforce. Keep your information secure by ensuring it doesn't leave your servers. All work is performed on your Remote Desktop which is located on the server, so data never needs to be downloaded to employee computers. Centrally manage application deployments and updates. Ensures all employees are using the same application versions and avoids individual device issues. VDC BlackSun Remote Desktop Services - Allow Desktop applications to be remotely accessed - Centrally manage application deployments and updates - Preserve user profiles settings across multiple servers - Support for multiple displays - GPU Pass-through for graphic intensive applications and 3D modeling - Improved speeds for low bandwidth connections - Easy Print and Printer Redirection
systems_science
https://www.chem.fsu.edu/bio-t.php?userID=1273
2017-04-26T04:03:00
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917121153.91/warc/CC-MAIN-20170423031201-00122-ip-10-145-167-34.ec2.internal.warc.gz
0.771357
357
CC-MAIN-2017-17
webtext-fineweb__CC-MAIN-2017-17__0__195022529
en
Our research focuses on determining the structures and interactions of macromolecules by ion mobility / mass spectrometry. We are interested in macromolecular systems of both industrial and biological importance. The two main areas of research are chemokine-mediated chemotaxis of immune cells and the structural determinants of non-conventional oils. The work is a blend of experiment and theory that brings a structural perspective to two of the Grand Challenges of the 21st century, health and energy. Y. Seo, A. Andaya, C. Bleiholder, J. A. Leary: Differentiation of CC vs CXC chemokine dimers with GAG octasaccharide binding partners: An ion mobility mass spectrometry approach. Journal of the American Chemical Society, accepted. C. Bleiholder, N. F. Dupuis, M. Murray Gessel, M. T. Bowers: Dimerization of chirally mutated Enkephalin neurotransmitters: implications for peptide and protein aggregation mechanisms. Journal of Physical Chemistry B 117, 1770-1779 (2013). C. Bleiholder, T. Wyttenbach, M. T. Bowers: A novel projection approximation algorithm for the fast and accurate computation of molecular collision cross sections (I). Method. International Journal of Mass Spectrometry 308, 1-10 (2011). C. Bleiholder, N. F. Dupuis, T. Wyttenbach, M. T. Bowers: Ion mobility–mass spectrometry reveals a conformational conversion from random assembly to -sheet in amyloid fibril formation. Nature Chemistry 3, 172-177 (2011).
systems_science
http://uk.iosafe.com/1517-nas-backup-archive-dr-appliance
2018-04-19T13:07:54
s3://commoncrawl/crawl-data/CC-MAIN-2018-17/segments/1524125936969.10/warc/CC-MAIN-20180419130550-20180419150550-00598.warc.gz
0.895773
733
CC-MAIN-2018-17
webtext-fineweb__CC-MAIN-2018-17__0__227524142
en
Powered by Synology® DiskStation Manager Scalable powerhouse for total data protection. The ioSafe 1517 is a fire- and waterproof 5-bay cloud-enabled NAS, engineered to protect business-critical data, enhance recovery times and deliver optimal reliability. Powered by Synology’s award-winning DiskStation Manager operating system, the 1517 NAS includes easy-to-use apps for collaboration, file syncing and sharing, video surveillance and more. Ultra-performance NAS optimized for intensive tasks, encryption and instant disaster recovery With its fire/water protection, superior performance, scalability, resilience, and comprehensive features, ioSafe® NAS 1517 is the ideal storage solution for your growing business. Featuring a quad-core processor, hardware encryption engine, and flawless software integration, the ioSafe 1517 is a powerful, expandable 5-bay NAS server that delivers outstanding speed and accelerated data encryption. The 1517 can help to simplify disaster recovery, business continuity, data management, sharing storage for storage and backup - with minimal setup and the freedom to expand capacity at any time. Ultimate Security and Protection The ioSafe NAS 1517 powered by Synology DSM is the ultimate in security because data remains local and is never exposed to the dangers on the traditional public cloud. All security details are 100% controlled by you – the person that cares the most about the data. Like an aircraft black box for data, there’s not another product on the market which combines the security, protection and features of the 1517. Be safe! Quad-core CPU for blazing-fast performance The 1517 is powered by a quad-core CPU running at 1.7 GHz. Combined with 2GB of DDR3 RAM, the 1517 is blazingly fast with 449 MB/s sequential writing performance. Fast encrypted data transmission. Encrypting data is essential to a healthy security policy. Equipped with a dedicated hardware encryption engine, the ioSafe 1517 is the ideal digital safe for sensitive data; files and folders are encrypted on the fly without any perceived performance loss. Quad LAN ports for failover and Link Aggregation support Equipped with four Ethernet ports featuring failover and Link Aggregation support, ioSafe 1517 ensures continual network uptime even if LAN connection malfunction occurs on one port, reducing the chance of service disruption and costly downtime. Link Aggregation improves connection speeds beyond the limits of a single network cable or port. Robust scalability up to 180TB As your data storage needs grow, ioSafe 1517 can be connected to dedicated expansion units for additional storage on the fly, growing capacity with minimal effort. With two expansion units connected through specially designed eSATA connectors, raw capacity can be as high as 180TB. Easy Backup Solutions To protect your critical business data, ioSafe 15177 can become a centralized backup target for consolidating fragmented and unstructured information across your network through a range of options. Synology's own backup wizard can back up information to another ioSafe NAS, Synology NAS or any rsync server, with the option of using an encrypted connection, or to an external hard drive via USB 3.0 or eSATA. The wizard can also back up your information to the cloud if you happen to have sufficient bandwidth, including Amazon® S3, Glacier, and STRATO HiDrive. 1 Performance figures could vary on different environments. 2 ioSafe NAS 1517 supports up to two ioSafe X517 Expansion Modules, sold separately. You're in good company with ioSafe
systems_science
http://computer-forensics.safemode.org/index487a.html?page=PC_Forensics
2020-01-22T07:33:41
s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579250606872.19/warc/CC-MAIN-20200122071919-20200122100919-00038.warc.gz
0.945464
381
CC-MAIN-2020-05
webtext-fineweb__CC-MAIN-2020-05__0__7997376
en
Personal Computer Forensics This page is for information relating to forensic investigation pertaining to PC technology. There are several key items to remember when performing a forensic investigation on a PC. Computer Forensics involves a great many steps but these are some basics to remember. |1.||Before removing or seizing a PC make sure to document all the connection and cables attached to the PC. This is important should your case go to trial. This is also helpful because it may identify other devices that are attached to the PC that could contain data such as external hard drives and PDA's.| |2.||If the PC is running at the time of seizer photograph what screens are open or what programs appear to be running on the PC. Again this could be important should your case go to trial.| |3.||If the PC is running should you power down the PC using the operating system? A lot will depend on the type of case you are working on. Using the operating system to power down the PC gives the operating system a chance to delete temporary file and make changes to date/time stamps. If in doubt unplug the PC from the power source. In many cases this will preserve the data environment as it was at the time power was removed.| |4.||Never turn on the PC without having proper write blocking devices or software in place. The simple act of turning on the PC can possibly alter critical data. Windows based operating system while booting up will alter many date/time stamps in the system as well as date/time stamps of documents.| |5.||Never work form the original data stored on the PC. Always make a forensic copy of the data and work from the copy. In fact making several forensic copies is recommended. Should something happen to the copy you are working with you still have another copy to work from.|
systems_science
https://midweststl.com/category/commercial-hvac-systems/
2024-04-20T20:14:53
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817674.12/warc/CC-MAIN-20240420184033-20240420214033-00566.warc.gz
0.927669
719
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__4057416
en
Industrial refrigeration systems are complex and critical systems used to preserve food, chemicals, and other perishable items. These systems rely on the proper functioning of a number of components and mechanisms, which must work together to maintain the desired temperature and humidity levels. The system’s performance can degrade over time. As a system degrades the following can happen: - Increased energy consumption - Decreased efficiency - Potential failure. It is essential to perform regular maintenance on industrial refrigeration systems to keep them in good working condition. Benefits of Regular Maintenance for Industrial Refrigeration Systems The benefits of regular maintenance of industrial refrigeration systems are many. Maintenance can help extend the lifespan of the equipment, thus saving costs that would have been incurred in replacement. Regular maintenance can also help identify small problems before they become large ones. This reduces the likelihood of downtime, which could have an adverse effect on business operations. Regular maintenance ensures that the equipment is running efficiently, which can lead to a reduction in energy costs. Condenser and Evaporator Coils in Industrial Refrigeration Maintenance of an industrial refrigeration system involves several tasks. One of the most important tasks is cleaning the condenser and evaporator coils. These coils are responsible for transferring heat and are therefore crucial to the system’s efficiency. Over time, dust, dirt, and other debris can accumulate on these coils, reducing their effectiveness. By cleaning these coils, the system can operate at maximum efficiency, saving energy costs and prolonging the lifespan of the system. Refrigerant Charge in Industrial Refrigeration Systems Another important maintenance task is ensuring that the refrigerant charge is correct. Refrigerant is used to transfer heat from one place to another, and it is vital to ensure that there is enough refrigerant in the system. Too little refrigerant can lead to inefficient cooling, while too much can damage the compressor. It is essential to check the refrigerant charge regularly and make any necessary adjustments. Compressor Check for Industrial Refrigeration Checking the compressor is also a vital maintenance task. The compressor is the heart of the refrigeration system, and it must be functioning correctly for the system to work efficiently. Regular inspections can detect any potential problems with the compressor, such as leaks or malfunctioning valves. Addressing these issues early can prevent costly repairs or replacements down the line. System Controls and Sensors in Industrial Refrigeration Another important maintenance task is ensuring that the system’s controls and sensors are working correctly. These controls and sensors help regulate the system’s temperature and humidity levels, and any malfunctioning can lead to inefficient operation. Regular testing and calibration of these devices can ensure that the system is working correctly and that the temperature and humidity levels are within the desired range. Benefits of Regular Maintenance in Industrial Refrigeration Regular maintenance of industrial refrigeration systems is essential to ensure that they are running efficiently and effectively. The benefits of regular maintenance include: - Increased equipment lifespan - Reduced energy costs - Reduced downtime. The maintenance tasks involved in industrial refrigeration system maintenance are varied and include cleaning the coils, checking the refrigerant charge, inspecting the compressor, and testing and calibrating the system’s controls and sensors. By keeping these components in good working order, businesses can ensure that their industrial refrigeration systems will continue to perform at a high level, protecting their products and their bottom line. Contact us at Midwest Mechanical Services and Solutions to plan for regular maintenance on your industrial refrigeration system.
systems_science
https://kiwigrid.com/en/article/kiwios-x-why-a-versatile-energy-iot-platform-is-the-key-to-a-successful-energy-transition-in-the-home
2023-09-26T09:35:58
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233510179.22/warc/CC-MAIN-20230926075508-20230926105508-00612.warc.gz
0.936253
1,325
CC-MAIN-2023-40
webtext-fineweb__CC-MAIN-2023-40__0__202937386
en
KiwiOS X: Why a versatile energy IoT platform is the key to a successful energy transition in the home The rapid expansion of renewable energies, combined with digitalization, is leading to a change in value creation within the energy market. A linear value chain with a one-way flow of energy from central energy producers to end customers has become a highly digitized energy system that places end customers at the center of value creation. The modern energy market is no longer just about the distribution and provision of energy but also the production, logistics and installation of hardware components such as PV systems, heat pumps and EV chargers. Digital platforms enable all these devices to connect with each other. Therefore, the home is becoming a major player in the energy transition, supporting numerous new business models. A cross-sector, robust and scalable energy IoT platform is needed to tap into this market. Kiwigrid's KiwiOS X platform is the ticket for anyone who wants to participate in the rapidly growing renewable energy market. The growing market for renewable energy in the home More than 100 million single and two-family homes in Western and Northern Europe, as well as hundreds of millions more worldwide, will contribute to the energy transition over the next 30 years. If 100 million private households invest €80,000 in PV systems, energy storage, EVs and heat pumps – all with an average lifespan of 20 years – and if average annual residual electricity costs and revenues from flexibility trading amount to €2,000 per year over the same period, the one-off purchase costs for hardware will result in a market worth €8 trillion. Additionally, the ongoing revenues will result in a market worth €4 trillion, which means that the energy transition in private households will represent a €12 trillion market1 in total. While mainly hardware manufacturers, wholesalers, distributors and installation companies are currently benefiting from the surge in demand for PV, heat pumps, storage systems and EVs, only companies that offer their customers holistic, cross-sector and complete solutions will be viable in the long run. As system integrators, these businesses are experts in PV, electricity and heat. They understand the importance of digitization and customer centricity and set up all processes accordingly in a way that is both mass scalable and easy to use. The energy market has numerous participants who either provide (supply) or purchase (demand) energy solutions. On the supply side, there are energy suppliers, hardware OEMs, complete system providers and installation companies. On the demand end, there are end customers, aggregators, network operators and balance responsible parties. Of all building types, the home is currently at the center of the modern energy market. As there are no complex ownership structures, investment decisions in decentralized energy solutions can be made and implemented quickly and easily. The home as the center of our planet's largest platform In private homes, heating and cooling, energy generation and storage, and mobility come together. PV systems are installed on roofs, EV chargers power EVs with self-generated electricity in garages, while heat pumps provide energy-efficient comfort all year round. The system is supported by home energy storage units. These energy devices already generate added value for the end customer, though they unleash much greater potential when connected to other devices. For example, a PV system can be connected to an energy storage device to store solar energy for sunless hours. A heat pump can also be operated with solar energy generated by the PV system. These links between devices open numerous new business models. The pure hardware market can be transformed into a service market. For end customers, this development means that they will no longer have to purchase each individual hardware component along with the necessary contracts and software services. Instead, they will be able to buy combined packages that support the intelligent connection of their energy devices. This requires an energy IoT platform with software capable of facilitating the complex connectivity of the numerous market players and their hardware solutions. Requirements for an energy IoT platform It is not enough to install a single gateway to generate convincing added value for end customers. In some cases, homeowners invest tens of thousands of euros in energy devices and expect competent and versatile energy management in return. To ensure this, the software solutions offered to end customers should be built on a robust energy IoT platform that meets the following requirements: - Easily maintainable Kiwigrid is already processing over seven billion data points daily from more than 220,000 connected energy devices in 16 European countries, ensuring a platform availability of at least 99.5 %. KiwiOS X, our operating system for renewable energy, uses specific APIs tailored to the device classes PV inverters, electricity storage, EV chargers, heat pumps and meters, regardless of device manufacturers and protocols. In the future, additional APIs will also connect smart tariffs and flexibilities. While specific APIs allow the modular implementation of solutions, the standardized, open API software architecture enables easy integration with adjacent systems. KiwiOS X proves its everyday suitability by interacting with our energy management gateway VoyagerX, which collects data points in the field and delivers them to the platform. As installation capacities are limited due to the shortage of skilled workers and rapidly increasing demand, installation must be as straightforward and fast as possible. The Energy Manager VoyagerX handles both of these tasks. KiwiOS X was fully migrated to the Google Cloud at the end of last year and uses Google Cloud services that comply with the C5 standard issued by the German Federal Office for Information Security. Kiwigrid's information security management system is also certified to the international ISO/IEC 27001 standard. An enormous complexity becomes easily accessible for companies The boundaries between formerly different sectors are becoming increasingly blurred: the solar, electricity, heat, mobility and IT sectors are increasingly converging. This convergence presents all market participants with the major challenge of adapting to a new market while generating value for their end customers. Many providers are still new to the energy market and are neither familiar with regulations nor existing software solutions. Therefore, the enormous complexity of an energy IoT platform is impossible to accomplish by individual energy market players alone. For this reason, Kiwigrid has focused on developing a platform that enables companies to connect various devices and systems without having to design their own software. Do you share our vision of bringing energy optimization, self-sufficiency and sector coupling to homes in a user-friendly way and on a large scale? Follow us on LinkedIn!
systems_science
http://lichteblau.blogspot.com/2007/03/klacks-parsing.html
2019-10-15T16:00:19
s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986660067.26/warc/CC-MAIN-20191015155056-20191015182556-00269.warc.gz
0.922456
549
CC-MAIN-2019-43
webtext-fineweb__CC-MAIN-2019-43__0__20885230
en
Starting with the current release of CXML, there is now a new parser interface called Klacks. Similar to StAX, the new interface is more convenient than SAX, while still providing the same features as the old one, including validation. Basically, the klacks parser can be used as a (rather sophisticated) tokenizer, and you get to write a recursive descent parser based on that. SAX and StAX are Java's protocols for XML parsing. Sometimes they are being referred to as low-level interfaces for "expert" use only (the suggested alternative being something like DOM), but their purpose is really to parse XML without building an in-memory representation. Low-level or not, they are the right choice when parsing into application-defined data structures or when performing simple on-the-fly transformation of XML data as it is being read. In SAX, an XML parser will process the entire document in one go, emitting events as it sees them. User code needs to implement its own handler class, with methods for the events it cares about. The SAX concept is known as "push-based". In contrast, the "pull-based" StAX parsing model is similar to working with an input stream. User code starts by creating an input stream object for the XML document, then reads events from that stream one by one. (Klacks uses the term source instead of stream, to avoid confusion with Common Lisp streams.) API design choices. StAX distinguishes between a high-level API, which creates a Java object for each event, and the low-level API, which just returns an enum indicating the type of event, and has separate methods to access the current event's data. Klacks has just one set of functions for both purposes, since it seemed more lispy to use multiple values. Instead of returning just a keyword indicating the event type, the main klacks functions always include useful event data as additional return values. Java's StAX also includes classes for XML serialization. No such extension was needed for CXML, since it already supports convenient serialization using SAX events. The with-element macro and related functions make generation of those events easy. Simple klacks parsing example: * (defparameter *source* (cxml:make-source "<example>text</example>")) * (klacks:peek-next *source*) :START-DOCUMENT * (klacks:peek-next *source*) :START-ELEMENT NIL ;namespace URI "example" ;local name "example" ;qualified name * ...
systems_science
https://flagma.com.br/en/industrial-water-treatment-equipment-o22756.html
2023-12-11T19:32:41
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679516047.98/warc/CC-MAIN-20231211174901-20231211204901-00052.warc.gz
0.866356
144
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__97141944
en
We offer high-quality equipment for water treatment Litech Aqua TM (produced in Ukraine) Using many years of practical experience, we can solve any task of water purification. We produce the following types of equipment: - Reverse osmosis systems; - Ultrafiltration systems; - Nanofiltration systems; - Microfiltration systems; - Water softening plants; - Deironing and de-watering plants; - Electrodialysis systems; - Electroionization systems; Each technological scheme of the equipment is calculated individually, based on the analysis of the source water and all individual customer requests. Each installation undergoes a 3D modeling process.
systems_science
https://www.gstarcad.uk/gstarcad-network-license-is-it-for-you/
2024-04-22T16:38:05
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296818312.80/warc/CC-MAIN-20240422144517-20240422174517-00867.warc.gz
0.878822
1,609
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__168106556
en
GstarCAD licensing is flexible and there are several types of licensing. This article will introduce the GstarCAD network license and how it works. All the advantages and disadvantages of using a network license will be presented. The article will help you decide to purchase a network license for GstarCAD. Types of licenses for GstarCAD Software First, the new user must decide which version of the GstarCAD he needs (Standard, Professional, Mechanical, Architecture). We advise the user which version based on his needs (workflow, projects, knowledge, etc) Then comes the question which type of GstarCAD licensing type to choose. The type of licensing depends mainly on the workflow and the number of customer users, there are several different licensing options available with GstarCAD. - Standalone Flexnet license(Standalone) - Network Flexnet license (Network) - Standalone Hardlock license (Standalone) on USB Dongle key - Network Hardlock license (Network) on USB Dongle k *Explanation of the terms mentioned above: – Flexnet means a ‘software’ license in the form of a serial number. – Hardlock indicates the type of license that is tied to the Hardware equipment – USB Dongle key in this case. – Network indicates a network – floating license, which is activated on the server – Standalone is an independent license (it is not connected to the local network and works ‘offline’) The type of license determines how the Software will be activated and how it will be used. We choose the type of licensing correctly based on customer’s workflow, the number of PC devices and the number of software users. For more information or answers to questions, please contact our CAD consultants. In the following, we will describe the GstarCAD network license. How does the GstarCAD network license work? The user must have a local network set up in the office, i.e. a server with a Windows operating system. On this PC or the License Manager application is installed on the server, it enables the activation and deactivation of the license and the management of the license data. The version of the Software and the number of seats the user owns are also displayed here. All GstarCAD users in the company install the GstarCAD software on their PC and enter the IP of the local network and the Port number (usually 27000) at the first activation. The user thus connects to the server, where the license seat will be ‘occupied’ when using the GstarCAD. If the company owns only 1 seat, only one person can use the license at a given time. After the user closes the program, another person from another computer can then use the program. Thus, it is practically an exchange of licenses between users. It is worth noting that the license should be deactivated in the License Manager application when there are changes or upgrades on the local server. After customer finishes on server maintenance or upgrade, then the license may be reactivated in the License Manager. This way the license will not get lost and the work will not be disrupted. For who is GstarCAD Network license suitable? The profile of eligible users for the network type license of the GstarCAD are: - Larger companies where control over program licenses is required and activation of standalone licenses would be time-consuming and complex - Companies where there are several users, but usually 1 person uses the program at a given moment For who is GstarCAD Network license NOT suitable? Unsuitable users for the network type license of the GstarCAD are: - Individual users who work independently and almost always have the same PC - Smaller companies that do not have an established local network (server) - Companies that do not want to have a license tied to the local network due to the nature of the work (they are often in the field or abroad) Advantages of GstarCAD Network license - easier management and control of a larger number of licenses – seats, all licenses are centralized at a place – a server, instead of on an individual computer. - An application is available for monitoring usage and recording licenses (Network Dashboard application) - the license can be used on any PC connected to the local network (of course if the GstarCAD is installed), - better exploitation of licenses – exchange of licenses between users - Savings: a more “Cost effective” solution when the company has several users of this program. - Flexibility of use: network licenses can be “borrowed” from the server and used anywhere outside the office. In this case, the license becomes ‘Standalone’ and works in ‘offline’ mode without a network. Cons of GstarCAD Network license - To work from home or from another location, a VPN interface is required or the ‘borrow’ license should be made. - The possibility of firewall or anti-virus problems, which may temporarily disable the use of licenses until the problem is resolved - The occasional possibility of intervention by these. support and IT department by the user if there are unexpected problems with the local network - In case of problems on the server (not functioning of this device) or network loss, the use of the program is disabled - Every time the server is tampered with (formatting, upgrades, etc.), the license must be deactivated and activated when the server is ready for use. - The server must always be running Min. System requirements The user must have a local network set up in the office, i.e. a server (server) with the Windows operating system. The License Manager application will be installed on the server computer. Also, every CAD user must install GstarCAD (on Windows platform). GstarCAD does not work on MAC and Linux operating systems. - OS (Operating System): Windows 11, Windows 10 version 1507 or higher: Home, Professional, Education, and Enterprise (LTSC and Windows 10 S are not supported), Windows 8.1 (with Update 2919355): Core, Professional, and Enterprise, Windows 7 SP1 (with latest Windows Updates): Home Premium, Professional, Enterprise, Ultimate, Windows Server 2019: Standard and Datacenter, Windows Server 2016: Standard and Datacenter, Windows Server 2012 R2 (with Update 2919355): Essentials, Standard, Datacenter - Note: Because GstarCAD and related tools are developed and compiled with visual studio 2017, in order to ensure the normal operation of the software, the operating system needs to be able to install the environment package of visual studio 2017 normally. - CPU: 1.6 GHz processor (3 GHz or faster Multi-core processor) - RAM: 2GB (8GB recommended or higher) - Graphics Card: 1 GB GPU (2 GB GPU recommended or higher) - Hard Disk: The system and installation disk need more than 1 GB of space. (4GB recommended or higher) - To improve performance, it is recommended that the software be installed and stored on a solid state drive (SSD). - Display Resolution: 1024*800 or above; GstarCAD will work best at a resolution of HD and 4K resolution. - Administrator rights are required to install GstarCAD and license tools. - .NET Framework 4.8 or above (Only for custom software development). - The network license server and all workstations running applications that rely on network licenses must support the TCP/IP protocol. Contact us for additional help and advice!
systems_science
http://oceancomputer.com/partners/drobo
2013-05-20T15:55:32
s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368699068791/warc/CC-MAIN-20130516101108-00062-ip-10-60-113-184.ec2.internal.warc.gz
0.880684
215
CC-MAIN-2013-20
webtext-fineweb__CC-MAIN-2013-20__0__134483479
en
Drobo makes award-winning data storage products for Small and Medium Businesses and Individual Professionals that provide an unprecedented combination of sophisticated data protection and management features, affordable capacity, and ease-of-use. With over 50 industry awards and hundreds of thousands of customers worldwide, Drobo has demonstrated success solving the three major storage challenges in one device – data protection, capacity adjustment, and application service-level optimization – through patented BeyondRAID™, Thin Provisioning / Reclamation, and the breakthrough Automated Data-Aware Tiering technology. Unlike other storage options, Drobo lets users hot-swap drives, mix and match drives of different capacities, speeds, and types (SATA / SAS / SSD), and perform zero-support drive pack migration. This means that customers can optimize storage economies without worrying about drive order, data loss, downtime, or compatibility. They can swap in drives as needed, while the Drobo is running. Copyright 2012. Ocean Computer Group Inc. | 90 Matawan Road | Matawan, NJ 07747
systems_science
https://abm.gov.pl/en/clinical-research/networks-of-clinical-tr/polish-clinical-trial-n/the-aim-and-main-assump/74,The-aim-and-main-assumptions.print
2021-12-07T20:30:18
s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964363418.83/warc/CC-MAIN-20211207201422-20211207231422-00324.warc.gz
0.943344
313
CC-MAIN-2021-49
webtext-fineweb__CC-MAIN-2021-49__0__75630664
en
The aim and main assumptions In Poland, apart of supporting the development of professional Clinical Trials Support Centers (CTSC), there is a need to create a network within would be possible to cooperate between the centers that create a network, consisting in the efficient exchange of information, experiences, development of IT tools dedicated to clinical trials, procedures and documents and management of studies. The purpose of establishing the Polish Clinical Trials Network (PCTN) is to implement uniform systemic quality and process solutions in all clinical trials centers in Poland. The implementation of the above solutions will have a direct impact on strengthening Poland's position in the field of clinical trials, increasing the attractiveness and competitiveness of the national infrastructure and its ability to support high-quality research, and consequently, will result in more effective development of cooperation at the international level. - developing a common quality standard - standardization of operational processes in the field of commercial clinical trials, effective implementation of non-commercial clinical trials, as well as early phase research with more efficient service - the opportunity to exchange knowledge and experiences, common solve problems of centers and improve national cooperation between centers in the network, which will significantly affect the process of effective recruitment on a national scale and improve the feasibility process - activities aimed at educating clinical trial staff and improving the qualifications of the personnel participating in a clinical trial - unifying the rules for conducting of research, contracting, valuation or settlements, in the long term will contribute to the optimization of costs in the field of reimbursement and drug policy in the health care system
systems_science
http://www.revnetmgt.com/what-we-do/hybrid-cloud-strategies-and-services/
2017-10-22T01:11:03
s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187824931.84/warc/CC-MAIN-20171022003552-20171022023552-00222.warc.gz
0.877554
441
CC-MAIN-2017-43
webtext-fineweb__CC-MAIN-2017-43__0__229922507
en
Hybrid Cloud Strategy A true Hybrid Cloud Strategy blends all the systems and services of Information Technology, strategically placing workloads where they provide the highest level of functionality. A hybrid cloud strategy must be started from the ground up, network, routing, services, then applications. Customers who combine RevNet’s managed express route, Hosting and Disaster Recovery solutions, and Microsoft Azure form a best of breed hybrid cloud solution! ExpressRoute Rapid Design ExpressRoute Rapid Design is a professional service offered by RevNet to discover and design an ExpressRoute connection(s) for our customers. This process involves a customer’s Network, Server, and Application teams to establish Layer 2 and 3 network design to effectively integrate ExpressRoute with the customer’s network. RevNet Architects help determine VLANs required, IP addressing schemes, VNET Gateway configurations, Layer 2 connection partners, etc… in this process. Determining the size of ExpressRoute connection required is an important part of an ExpressRoute implementation. Through the Rapid Design process we help determine usages of applications and services in both Microsoft Azure and other Microsoft Online services such as Office 365. Then we help determine the required bandwidth utilization for all services combined. All of this data is then rolled up in to a rapid design package, ready for implementation! RevNet offers a unique Managed ExpressRoute service that is not telco dependent. This service helps accelerate ExpressRoute deployments over standard telco based deployments. ExpressRoute presents challenges as it blends cloud based networking and architecture, with BGP peering, QinQ VLANs, and complex routing schemes. RevNet’s years of experience in hosting services, complex Internet routing, and now public cloud infrastructure, we can provide all the skillsets required to make your Hybrid Cloud happen fast! The Managed ExpressRoute service from RevNet is also uniquely positioned to help customers with multiple hybrid cloud needs. From on premises hosted applications and systems to Disaster Recovery, VM hosting and physical colocation – to public cloud workloads – our Managed ExpressRoute service provides a strategic platform for all our customer’s Hybrid Cloud needs. Our Managed ExpressRoute Service features the following:
systems_science
http://intervenmedical.com.br/solucoes/videolaparoscopia/abcflex-argon-beam/
2020-06-01T07:22:29
s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590347415315.43/warc/CC-MAIN-20200601071242-20200601101242-00007.warc.gz
0.851322
229
CC-MAIN-2020-24
webtext-fineweb__CC-MAIN-2020-24__0__96330362
en
ABCFlex Argon Beam Argon Beam Coagulation Advanced technology provides non-contact hemostasis with minimal thermal injury of adjacent or underlying tissue. ConMed provides advanced technology for the gastroenterologist with ABCFlex™ Probes and the System 7550™ ABC® ESU or System 7500® ABC® ESU. Argon Beam Coagulation® technology enables the surgeon to achieve non-contact hemostasis with minimal thermal injury of adjacent or underlying tissue. ABC® technology is ideally suited for GI applications and is a valuable tool for the gastroenterologist. Like standard electrosurgery, energy is delivered to tissue to create hemostasis, but the argon beam uses a focused column of ionized argon gas at room temperature to deliver the energy. The resulting cool argon beam moves blood away from the bleeding tissue and coagulates directly to the mucosal surface which is intended to limit thermal spread and may result in less necrosis. It can be used to create superficial ablation of mucosal tissue over a large area, or remedy Arteriovenous Malformations (AVMs).
systems_science
https://kontaksupport.zendesk.com/hc/en-us/articles/16187569067540-Grandstream-HT812-Quick-Start-Guide
2023-12-05T12:16:57
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100551.17/warc/CC-MAIN-20231205105136-20231205135136-00708.warc.gz
0.873752
224
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__26106165
en
Setting up a Grandstream HT812 ATA (Analog Telephone Adaptor). - Connect the power cord the port labeled “DC 12V”. - Connect an internet source to the ethernet port with the “Globe” Icon. - Connect an analog phone line from the fax machine directly to either Line 1 or Line 2. WARNING: Do not connect an analog phone line from a wall phone jack to either Line 1 or Line 2 as doing so could cause damage the device. When the device first starts up, the power icon will flash. Once booted, the power icon will stay solid. A solid internet connection is indicated by Globe icon which should remain solid and blink occasionally. Line 1/Line 2 will flash during an Incoming and outgoing fax. If Line 1/Line 2 blinks continuously without stopping, there may be a problem with the ATA. The Pass-Through Port, indicated by the network tree icon, allows an internet connection to be shared with another device in the event that there are not enough Ethernet ports to go around.
systems_science
https://www.ecoppia.com/media/press?page=4
2024-02-26T01:31:38
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474649.44/warc/CC-MAIN-20240225234904-20240226024904-00775.warc.gz
0.897785
271
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__113382723
en
Real-time and forecasted weather intelligence integrated into E4 control system enables more efficient plant operations and higher energy output The robotic cleaning solution is the first and only system verified for long-term use on nine major solar panel brands Ecoppia’s robotic water-free solution has cleaned over 3 million panels to-date. Already commercially deployed, the Ecoppia E4 system helps solar parks deliver more energy, more efficiently. Ecoppia, an innovative developer of autonomous water-free solar panel cleaning solutions, announced today that JA Solar, a leading manufacturer of photovoltaic solar panels, has certified Ecoppia’s solution for use on its solar panels. Ecoppia recently announced similar certification by Suntech Power. Additional certification processes are underway, with the aim of ensuring Ecoppia’s market–wide compliance with the stringent standards of all major panel manufacturers. “Ketura Sun” solar park in Israel’s Negev desert is the first utility-scale solar energy production facility to implement an autonomous, water-free robotic panel cleaning system – Ecoppia’s E4. Commercially deployed on Suntech panels, Ecoppia’s automatic solar array cleaning solution optimizes solar power generation by increasing production and lowering costs.
systems_science
https://www.questys.com/business-systems-integration/what-is-document-management/
2024-02-22T17:42:24
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947473824.13/warc/CC-MAIN-20240222161802-20240222191802-00386.warc.gz
0.870492
273
CC-MAIN-2024-10
webtext-fineweb__CC-MAIN-2024-10__0__74270066
en
Enterprise Content Management (ECM) is a set of defined processes used to obtain, organize, store and deliver unstructured content crucial to an organizations operation in the most effective manner possible. The primary goal is to enable organizations to deliver relevant content to users where and when they need it as well as eliminate ad hoc processes that can expose an organization to regulatory compliance risks and security issues. How can an Enterprise Content Management (ECM) system help your organization? Capture, manage, store, preserve, and deliver content and documents related to organizational processes electronically. Features include document and records management, auditing capabilities, check – in/check – out and other security mechanisms. Streamline the legislative agenda process from start to finish. Automate the agenda process of creating staff reports, agendas, and packets; real – time roll – call, vote tabulation, meeting minutes, agenda archival, backup material, keyword search and retrieval. Who benefits from Enterprise Content Management? Ease of Deployment & Support Questys is designed for rapid deployment without any specialized, pre – existing knowledge. Business systems can be integrated without using custom code. Files are secured allowing only those with the proper security clearance to access sensitive information. Cost Savings, Efficiency & Compliance ECM can save any organization thousands of dollars in operational costs.
systems_science
https://www.gouspack.com/solutions/
2020-08-04T22:53:14
s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439735885.72/warc/CC-MAIN-20200804220455-20200805010455-00283.warc.gz
0.881119
123
CC-MAIN-2020-34
webtext-fineweb__CC-MAIN-2020-34__0__2943897
en
ADDI is the technology brain behind USPack’s Adaptive Delivery Solutions model. This fully-integrated suite of technology tools combines information portals, scheduling software, and real-time intelligence data to inform and enhance delivery execution and customer experiences across the delivery spectrum. This flexible platform delivers unmatched levels of transparency, intelligence and predictability to our customers and their customers. From real-time tracking tools to high-level trend data and more, ADDI consistently delivers actionable insights that allows us to adapt quickly to changing needs and optimize our delivery solutions. Learn how ADDI can make your delivery operation smarter.
systems_science
http://buy-hosting.net/web-hosting-articles/cpanel-is-the-best-web-hosting-control-panel.html
2017-02-22T08:22:23
s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501170925.44/warc/CC-MAIN-20170219104610-00603-ip-10-171-10-108.ec2.internal.warc.gz
0.904366
845
CC-MAIN-2017-09
webtext-fineweb__CC-MAIN-2017-09__0__36048742
en
Top 10 Reasons Why cPanel is the Best Web Hosting Control Panel Short answer: It rocks! 17 Sep 2014, posted by Michael Lederstatter 1. Easy Installation Unlike other web hosting services that require complex setup conditions, the cPanel only requires a LINUX environment for installation and setup. The manufacturers of cPanel have already programmed cPanel to incorporate installation features. Users can install and run cPanel with minimum requirements including 512 MB RAM, 266 MHZ and available disk space of 10 GB. 2. Versatile email control Users of cPanel can exercise maximum control over email hosting incluidng features like adding email accounts, deletion, editing and forwarding among others. Other important cPanel email features include specifying email box sizes, configuring email clients and customized autoresponders. Advanced features include 'SpamAssasin' that deletes any unwanted emails or categorizes spam mails into 'white lists' and black lists'. The user can also set email filters to customized settings for easier manageability. Security is the biggest worry experienced by any user including veteran programmers. Security benefits offered by cPanel include 'Leech Protection'; a feature that protects secure areas of a website from outside infilterators. Other security features include the 'IP Deny Manager' that prevents unauthorized users from accessing single or multiple IP addresses. 4. SSH/Shell Security One of the most delicate operations online is the sending of files containing confidential information. Users can activate the SSH/Shell (SSH stands for Secure Shell that typically connects the port 22 site on two computers to effect a transfer) security feature that allows for secure transfer of files over the Internet. In addition, the GNUPG (GNU licensed Public Guard Software) keys allow public encryption of private areas of a webiste for a limited amount of time. This time limit can be set by the user. For users who own and operate ecommerce stores, cPanel offers what it calls the SSL/TLS security service. SSl stands for Secure Sockets Layer and TLS is Transport Layer Security. Both offer extra protection against theft of sensitive information (including bank and payment information) online. 5. File Management Users gain access to automatic backup services that save entire directories including MySQL, home directories as well as email filters etc. In the event of a power failure or a website crash, the user has the option of uploading these important files from the backup version. 6. Legacy Manager and File Manager Legacy Manager that offers file access and functionality from any location. In other words users can move, copy, delete edit or create files from any location as long they use the correct passord and login information. The 'File Manager' is a more user-friendly variant of the Legacy Manager and offers an easier interface including the WYSIWYG HMTL editor. 7. Domain Management Users can create and view multiple sub-domains simultaneously as well as modify web-pages without having to make changes to htaccess files. In addition, cPanel offers add-on domains that can be linked to a single hosting account. The users can link 'parked domains' (unused domains) to their primary domain through the use of DNS servers. 8. Friendly User Interface The biggest plus point of cPanel is that it offers a simple and intuitive user interface that uses graphics and colors to make navigation and understanding easy, even for those who do not possess technical knowledge about web hosting. 9. Integration features Users can choose to integrate cPanel with an auto-installer like Fantastico. Fantastico is able to support more than 50 open source scripts. Using open source scripts saves time and effort and the user can create or make changes to web-pages with a few simple clicks. Some important script features include the ability to suspend or delete email accounts and also the option to restart the server over SSH. 10. VPS Optimization Users gain access to an impressive array of VPS (Virtual Private Server) optimization features. Users typically face problems of slow website or application functioning during times when there is high competition for resources.
systems_science
https://technologytickles.com/page/16/
2023-06-02T09:39:00
s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224648465.70/warc/CC-MAIN-20230602072202-20230602102202-00421.warc.gz
0.907813
547
CC-MAIN-2023-23
webtext-fineweb__CC-MAIN-2023-23__0__109304347
en
HashiCorp is a Software firm who are now very familiar to the IT world. Thanks to Terraform! What is Consul? Consul is a Service discovery and KV store developed by this company. It’s high availability architecture, security and the simple GUI makes it a point to look at this product. How does this work? Consul consists of two types of agents, client and server. Client consul agents are installed on the servers where we want to monitor the services on that server using consul service discovery feature. Server consul agents are the actual consul boxes. Let’s see a sample architecture and get into the details – The above architecture is from HashiCorp’s website. Service Discovery – As you see in the above architecture that the vault servers we have consists of consul clients installed on them. These consul clients sends the status of the services on the vault servers to the consul servers. From the consul server UI you can view the service status of those vault servers. Key Value Storage – Whenever you store any KV pair in the vault server, you are using consul backend to store the data. The consul clients sends the KV pair to the consul server and later, the consul server stores the data in encrypted format. When you try to read the KV pair on the consul server, you can see that the data is in encrypted format. To view the data in decrypted format you should be using vault server. Consul Server working model – These servers follow ‘leader and follower’ working model. The above architecture has a fault tolerance of ‘1’. What does this even mean? So, if one of the consul server fails among the 3 servers the other 2 servers are able to function normally without any glitches. As soon as the third one comes online it is able to automatically join the consul cluster. When you spin up 3 consul servers in the consul cluster, consul automatically elects one of them as the leader or the active server and the other 2 are followers or read replicas. There is no need to configure a load balancer as this is taken care by consul itself. Sounds simple right?!. - Consul’s architecture is highly available. - Fault tolerant. - Data is encrypted. - Simple open source product with an open source UI. - HashiCorp provides support for enterprise versions. - Easy to use. - Most importantly, we can act quickly when data is compromised with consul and secure the data.
systems_science
https://crictech1.com/52818/
2023-03-22T01:00:56
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296943747.51/warc/CC-MAIN-20230321225117-20230322015117-00679.warc.gz
0.926274
641
CC-MAIN-2023-14
webtext-fineweb__CC-MAIN-2023-14__0__109850151
en
Global economic uncertainty, climate commitments and digital transformation were three topics that headlined at the World Economic Forum in Davos recently. As global trade continues to slow, the Ukraine conflict and its far-reaching economic, people and political consequences have replaced the pandemic as the primary drag on western, globalised economies. As rising trade tensions and economic nationalism continue to influence world trade relationships, countries and organisations are adjusting to these changing geopolitical and economic dynamics. And, while it’s still too early in the year to determine whether the global economy will thrive or decline in 2023, one thing is clear: to circumnavigate the challenges ahead, it will require strategies and processes that are based on resilience and agility as businesses reassess their traditional approaches to business models and playbooks. As organisations continue to diversify their trading and manufacturing relationships to reduce global investment risks, building resilience and agility into supply chains will become a significant priority once more. Recent trade patterns provide clear evidence that many organisations across a variety of verticals are prioritising supply chain resilience, diversification and agility, particularly due to inflationary pressure and the cost of living crisis. Below are five key areas all retail and supply chain directors need to address if they are to insulate their brands against the chill of economic uncertainty in 2023: - Increase resilience by investing in technologies that provide complete and real-time visibility of inventory across all digital and physical sites – if you’ve got it, you should be able to sell it - Fund R&D initiatives to identify alternative approaches and drive innovation. When others are battening down the hatches, it’s often the best time to double-down on innovation and investment, enabling you to steal a march on your competitors when fairer winds return and consumer expectations have changed - Use modern tech, such as cloud-native, unified supply chain platforms or digital-twins to leverage the ‘force-multiplier’ effect across your inbound and outbound flows. These types of technologies (independently or collectively) can improve the early visibility of developing shocks and disruptions to deliver alternative scenarios, contingency plans and drive overall supply chain resilience - Accelerate efforts to incorporate scenario planning into capital allocation and supply chain management processes. Coupled with the right technology, scenario planning can improve long-term supply chain resilience and overall growth strategies. - Recognise and act on the popular zeitgeist of the age. It’s not enough to emphasise environmental, social, and governance objectives (which have implications for global supply chains) to your customers and the end consumers, it’s as important to ‘live’ these values with your employees too. In the face of technical complexity and macroeconomic, geopolitical volatility, leaders must build more resilience into their daily processes for everything from people management to supply chains. Being ready to turn adversity into opportunity has always been the mantra of successful businesses around the globe. However, in an age when there are more challenges (business and societal, technology and environmental), nuances and complexities than ever before, these five guidelines should help organisations navigate possible choppy waters ahead.
systems_science
https://starcite.smarteventscloud.com/rsvp/invitation/invitation.asp?id=/m2faf095-7JQGMGUF0J8P
2020-07-05T06:47:13
s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655887046.62/warc/CC-MAIN-20200705055259-20200705085259-00104.warc.gz
0.878171
456
CC-MAIN-2020-29
webtext-fineweb__CC-MAIN-2020-29__0__58274631
en
Conference “Big Data Analytics in the Life Sciences Industry” June 1-2, 2015 / Diegem (Brussels), Belgium Organized by the ExaScience Life Lab* * funded by IWT, Janssen Pharmaceutica, Intel The conference will bring together international experts in the area of Computation and Life Sciences. It is a forum to discuss the state-of-the-art of the hot topics in the rapidly growing area of Big Data Analytics in the Life Sciences Industry. The volume and the diversity of the data available to the life sciences industry today is such that automated approaches are quickly becoming of strategic importance. Think about screening 7.5 million images for activity of compounds. Or how to help drug developers select which of 2.5 million compounds would be worth investigating in more detail ? Or how to find markers in the vast amount of measurements that are becoming available to enable early detection for disease interception in Alzheimer's ? The increased compute capabilities, puts Machine Learning at the forefront of such automated approaches to tackle these questions. Day two of the event will focus on this topic more in depth with key speakers in the field. The ExaScience Life Lab is a cooperation between Janssen Pharmaceutica, imec, Intel and the Flemish universities: KU Leuven, U Gent, U Antwerp, VU Brussels, U Hasselt. It innovates Big Data Analytics in the Life Sciences Industry. During the conference also the main results of this collaboration will be presented. Keynotes are confirmed from - Georgia Papathomas (VP CIO Pharma, J&J World Headquarters US) - Mark Landy (IT IS VP, Enterprise Architecture, J&J World Headquarters US) - Simon Lovestone (Professor of Translational Neuroscience at University of Oxford; Leader of the NIHR Translational Research Collaboration in Dementia) - Samuel Kaski (Director of Helsinki Institute for Information Technology HIIT and Finnish Centre of Excellence in Computational Inference Research COIN at Aalto University and University of Helsinki) The agenda will be sent out in the next weeks. Attendance is by invitation only.
systems_science
http://www.regenapps.com/
2017-04-29T17:23:15
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917123549.87/warc/CC-MAIN-20170423031203-00187-ip-10-145-167-34.ec2.internal.warc.gz
0.948355
236
CC-MAIN-2017-17
webtext-fineweb__CC-MAIN-2017-17__0__27900432
en
RegenApps excels in providing industry specific and niche technology solutions for businesses. Our services include process and systems design, package implementation, custom development, business intelligence and reporting, systems integration, as well as testing, maintenance and support. RegenApps's expertise spans across the entire software technology spectrum. RegenApps has the unique capability to serve as the one-stop shop solutions provider for its customers. Our ability in being a complete solutions provider removes the hassle of managing multiple vendors and technological bottlenecks. It gives our customers the peace of mind that they have chosen a reliable, long-term partner capable of solving their development needs as their businesses grow. RegenApps understands that when it comes to technology solutions, one size does not fit all. Our approach provides the maximum value for the investment. Our team proactively suggest improvements as well as newer tools and technologies that can be used by our customers to reduce costs and make business systems more efficient. Our solutions are catered for the long-term benefits of our customers, balancing the needs of individual organizations while keeping an eye on the fiscal estimate. OUR MOBILE APPS DEVELOPMENT EXPERTISE
systems_science
http://riainvision.com/refphys/security.aspx
2017-03-30T10:41:56
s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218193716.70/warc/CC-MAIN-20170322212953-00274-ip-10-233-31-227.ec2.internal.warc.gz
0.910191
372
CC-MAIN-2017-13
webtext-fineweb__CC-MAIN-2017-13__0__317663201
en
Invision Sally Jobe is dedicated to maintaining HIPAA compliance, data security, and a protected and reliable technology infrastructure. HIPAA Compliance and Security Our technology and image delivery processes are HIPAA compliant. We designed our processes and technical infrastructure to guarantee secure and confidential data transactions to and from anywhere in the world. Our PACS system has strong security in which functionality for access tracking, restricting and/or limiting access to individual users, multiple users, and groups are centrally administered. For example, a referring physician may be restricted to only view patient studies and reports that he\she has referred to Invision Sally Jobe. Our security methodology is based on a "best practice" approach derived from industry standards. Secure Imaging Distribution The InteleViewer uses Secure Socket Layer technology with 128-bit DES encryption to create a secure connection to images and reports over the Internet. Additionally, a system user must login to InteleViewer with an assigned username and password. This ensures that only authorized users may view exams and reports and enables full auditing for maintaining HIPAA compliance throughout the entire process. The Invision Sally Jobe network and data center are supported 24 hours a day, 7 days a week, 365 days a year by dedicated staff to make sure data is accessible. The Invision Sally Jobe data center is located in Englewood, Colorado in a secure building with 24/7 FirstWatch monitoring. Inside, advanced environmental and security system monitoring ensure your data is protected and secure. Digital cardkey access and access code locks provide multiple levels of protection from unauthorized access. Invision Sally Jobe utilizes a multi-data center approach to provide 7x24 redundancy and access to images and reports. Each data center has an advanced fire suppression system which minimizes damage and doesn't require costly and time-consuming cleanup.
systems_science
http://softwaredownloaddatabase.info/download/ccleaner/
2017-04-29T11:29:03
s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917123491.68/warc/CC-MAIN-20170423031203-00248-ip-10-145-167-34.ec2.internal.warc.gz
0.92653
400
CC-MAIN-2017-17
webtext-fineweb__CC-MAIN-2017-17__0__148234548
en
There’s a reason why millions of PC users flock to CCleaner when they are trying to optimize their PC’s performance. It’s easy to use, powerful, and efficient at cleaning out your computer’s filesystem and registry of all the junk files and errors. On top of it all is a feature rich GUI that allows you to isolate the clutter in your system for a thoroughly customizable clean every single time. If you computer is running slow, it may not come as a surprise. As your computer runs, it collects mounds of junk files and registry errors that make it run slow and sometimes even crash. CCleaner knows exactly where to search for these files when it analyzes your system, listing all of the possibilities for freeing up space. When you browse the internet, advertisements save tracking cookies on your computer that allow them to monitor your browsing habits and target advertisements tailored to you. CCleaner can clean these cookies off of your system as well, leaving your browsing habits unmonitored and private. How it works When your computer starts up, programs tend to silently start up. CCleaner allows users to disable startup applications so that your computer boots up faster. Then, hackers and malicious attackers are always trying to gain access to your system, and deleted apps often leave files on your computer long after they are uninstalled, leaving fragmented files and broken registry entries that can wreak more havoc on your computer. By performing regular maintainance with this software, your computer can remain optimized and clean. With CCleaner scrubbing even the dark corners of your computer, you can bet that it gets everything. When opened, users are presented with an easy to use user interface that anyone can learn with a few minutes of dedication. Using a tab view, users can clean their filesystem of any temporary or junk files, clean brken registry entries, and analyze their system for performance info related to CCleaner’s tools.
systems_science
https://lbbusinessjournal.com/aerospace/rocket-lab-delivers-demonstration-satellite-to-orbit-for-space-force/
2023-12-02T22:35:21
s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100452.79/warc/CC-MAIN-20231202203800-20231202233800-00013.warc.gz
0.915691
350
CC-MAIN-2023-50
webtext-fineweb__CC-MAIN-2023-50__0__178693144
en
Rocket Lab successfully delivered a U.S. Space Force satellite to low Earth orbit late Wednesday evening, local time, the company announced Thursday morning. The Long Beach-based rocket manufacturer and launch service provider’s Electron rocket blasted off from its New Zealand complex at 6 p.m. Thursday evening (11 p.m. PDT Wednesday) carrying a demonstration satellite called Monolith. The mission is the company’s second for the USSF. “Congratulations to all the teams behind Monolith,” Rocket Lab founder and CEO Peter Beck said in a statement. “We’re proud to have safely delivered another mission to orbit for the United States Space Force.” Monolith will demonstrate the use of large deployable sensors to determine if they disrupt the satellite’s attitude, or orientation in space. The satellite also will serve as a platform to test future space protection capabilities. The mission was procured by the U.S. Department of Defense Space Test and Rocket Systems Launch programs, both of which are based at Kirtland Air Force Base in New Mexico. Dubbed “It’s a Little Chile Up Here,” the mission name pays homage to the state’s famed green chiles. “Programs like the Rapid Agile Launch Initiative are shining a light on the crucial role small launch can play in supporting fast-paced innovation in orbit to support innovation and space capabilities,” Beck said. “It’s a Little Chile Up Here” is Rocket Lab’s 21st Electron rocket launch since the company was founded in 2006. It comes on the heels of a rare failed mission on May 15.
systems_science
https://blog.baileighindustrial.co.uk/largest-cnc-benders-scandinavia/
2022-05-24T00:10:01
s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662562106.58/warc/CC-MAIN-20220523224456-20220524014456-00133.warc.gz
0.8876
140
CC-MAIN-2022-21
webtext-fineweb__CC-MAIN-2022-21__0__33307781
en
This enormous MB-220 CNC can handle 8” schedule 40 pipe, and is designed to operate 3 shifts a day, 365 days a year. Installed to aid in the manufacturing of ships, this fully automated system completely removes the “human factor” from the manufacturing process. The easy to use touchscreen programmer stores thousands of programs with multiple bends per piece. Baileigh Industrial is leading the world in industrial machinery installations. Many machines are in stock and ready to ship from multiple warehouses around the world. Call our team or email [email protected] to schedule a live demonstration.
systems_science
https://srijitmitra.net/2020/07/24/eks-custom-ami-blog/
2022-01-27T16:36:52
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320305277.88/warc/CC-MAIN-20220127163150-20220127193150-00181.warc.gz
0.931935
137
CC-MAIN-2022-05
webtext-fineweb__CC-MAIN-2022-05__0__230640567
en
This blog post walks through the steps to provision a Kubernetes cluster on AWS EKS using infrastructre as code. The primary use case is customers who have strict requirements around the machine images they are allowed to use in their organization. This solution deploys self-managed worker nodes to give the user the flexibility to choose their own custom machine images. It provides all the automation (including a pipeline) to spin up multiple worker nodes and migrate live workloads from one to another. The entire solution is built using AWS CDK and has been open sourced. Source code: https://github.com/aws-samples/eks-ami-tester
systems_science
http://braun.gatech.edu/
2015-10-07T20:59:56
s3://commoncrawl/crawl-data/CC-MAIN-2015-40/segments/1443737893676.56/warc/CC-MAIN-20151001221813-00146-ip-10-137-6-227.ec2.internal.warc.gz
0.93759
1,164
CC-MAIN-2015-40
webtext-fineweb__CC-MAIN-2015-40__0__115280602
en
Dr. Robert D. Braun has over 25 years experience performing design and analysis of planetary exploration systems as a member of the technical staff of the NASA Langley Research Center and the Georgia Institute of Technology. His research has focused on systems’ aspects of planetary exploration, where he contributed to the design, development, test and operation of several robotic space flight systems. He has also served as a leader and senior manager for a number of large, diverse engineering organizations at NASA. In 2010-2011, he served as the first NASA Chief Technologist in more than a decade. In this capacity, he served as the senior Agency executive for technology and innovation policy and programs. He created and led the initial implementation of a spectrum of broadly applicable technology programs designed to build the capabilities required for our nation’s future space missions. This activity spanned all ten NASA Centers, industry and academia, and included building technology partnerships between NASA and other government agencies. Dr. Braun successfully advocated for an increased NASA technology budget in a time in which the fiscal environment required the overall Agency budget to decrease. He also created, staffed and managed the NASA Office of the Chief Technologist and worked to improve NASA’s organizational culture towards one that values technology development and high-risk, high-reward research. Dr. Braun co-founded and serves as Chief Technology Officer of Terminal Velocity Aerospace, LLC, a small business providing atmospheric reentry services to enhance safety and promote the utilization of space. TVA offers small spaceflight systems designed to provide unprecedented data on the physics of reentry breakup, and for the safe return of small payloads from space. Dr. Braun joined the Georgia Institute of Technology as the David and Andrew Lewis Associate Professor of Space Technology in 2003. He was promoted to the rank of Professor in 2009. He leads an active research and educational program focused on the design of advanced flight systems and technologies for planetary exploration. This research integrates aspects of conceptual design and analysis, technology development, computational modeling and simulation and experimental validation. Recent research projects include the development of entry, descent and landing technologies for human Mars exploration, design and analyses of inflatable aerodynamic decelerators, pinpoint landing technology assessment for planetary exploration systems, and engineering strategies for asteroid deflection and orbital debris prediction. He is responsible for undergraduate and graduate instruction in space systems design, astrodynamics and planetary entry. Dr. Braun has led the design and technology maturation of multiple space systems. He was a member of the Mars Pathfinder design and landing operations team from 1992 to 1997 and has been part of development teams for the Mars Microprobe, Mars Sample Return and Mars Surveyor 2001 projects. From 1998-2000, he managed the development of the Mars Sample Return Earth Entry Vehicle, an innovative, risk-based entry system design. From 2001-2003, he served as the Mission Architect and Atmospheric Flight System Manager for the Aerial Regional scale Environmental Survey Mars Scout mission, a proposed scientific survey using a Mars airplane. In this capacity, he managed the Mars airplane development including the successful ground-based and high-altitude flight test program. Since 2005, he has focused on the technology maturation of a wide range of entry, descent and landing technologies, including single-stage entry systems, pinpoint-landing guidance strategies, inflatable aerodynamic decelerators and supersonic retropropulsion. He has been an active participant in the development of advanced methods for multidisciplinary design and optimization. Dr. Braun developed the Collaborative Optimization architecture while at Stanford from 1991-1996. This architecture was shown to have significant computational and operational benefits in the optimization of large, distributed design problems. Since completing the initial research in this area, several university and industry groups have applied this technique in solving a diverse set of engineering challenges. From 2000-2001, he led and integrated NASA’s advanced engineering environment development program. Dr. Braun has provided independent assessment and served on NASA review boards for Mars Polar Lander, Mars Odyssey, Mars Exploration Rover, Phoenix Mars Scout, Genesis, Mars Science Laboratory, Mars 2020, and the Orion Crew Exploration Vehicle. He has also served on multiple NRC Committees. Dr. Braun was invited to serve as a Moore Distinguished Scholar at the California Institute of Technology from January through June 2015. He has both teaching and research responsibilities while at Caltech. While in Pasadena, he is also collaborating with engineers and scientists at the Jet Propulsion Laboratory. Dr. Braun received a B.S. in Aerospace Engineering from Penn State in 1987, M.S. in Astronautics from the George Washington University in 1989, and Ph.D. in Aeronautics and Astronautics from Stanford University in 1996. He has received the inaugural AAS Space Technology Award (2014), 2012 Alvin Seiff Memorial Award, the 2011 AIAA von Karman Astronautics Award, 1999 AIAA Lawrence Sperry Award, the NASA Distinguished Service Medal, two NASA Exceptional Achievement Medals, two NASA Inventions and Contributions Team Awards, and nine NASA Group Achievement Awards. He is a member of the National Academy of Engineering, Vice Chair of the NRC Space Studies Board, Editor-In-Chief of the AIAA Journal of Spacecraft and Rockets, an AIAA Fellow, and the author or co-author of over 275 technical publications in the fields of atmospheric flight dynamics, planetary exploration, multidisciplinary design optimization, and systems engineering. He presently serves on Advisory Boards for the Jet Propulsion Laboratory, the Space Systems Sector of the Charles Stark Draper Laboratory and the Planetary Society. He is married to Karen G. Braun, is the proud father of Zack, Allie and Jessica Braun, and resides on a small farm in Newnan Georgia.
systems_science
https://www.12volt.com.au/victron-multiplus-ii-48-3000-35-32-inverter-charge
2024-04-17T18:41:23
s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817171.53/warc/CC-MAIN-20240417173445-20240417203445-00646.warc.gz
0.797155
1,786
CC-MAIN-2024-18
webtext-fineweb__CC-MAIN-2024-18__0__48093094
en
Victron MultiPlus-II 48/3000/35-32 Inverter Charger Victron MultiPlus-II 48/3000/35-32 Inverter Charger The MultiPlus-II, as the name suggests, is a combined inverter and charger in one elegant package. Its many features include a true sine wave inverter, adaptive charging, hybrid PowerAssist technology, plus multiple system integration features. MultiPlus Inverter/Charger - Lithium Ion battery compatible. The MultiPlus-II combines the functions of the MultiPlus and the MultiGrid. It has all the features of the MultiPlus, plus an external current transformer option to implement PowerControl and PowerAssist and to optimize self-consumption with external current sensing (max. 32A). It also has all the features of the MultiGrid with built-in anti-islanding and an increasingly long list of country approvals. A MultiPlus, plus ESS (Energy Storage System) functionality The MultiPlus-II is a multifunctional inverter/charger with all the features of the MultiPlus, plus an external current sensor option which extends the PowerControl and PowerAssist function to 50A resp. 100A. The MultiPlus-II is ideally suited for professional marine, yachting, vehicle and land based off-grid applications. It also has built-in anti-islanding functionality, and an increasingly long list of country approvals for ESS application. Several system configurations are possible. For more detailed information see the ESS Design and configuration manual. PowerControl and PowerAssist - Boosting the capacity of the grid or a generator A maximum grid or generator current can be set. The MultiPlus-II will then take account of other AC loads and use whatever is extra for battery charging, thus preventing the generator or grid from being overloaded (PowerControl function). PowerAssist takes the principle of PowerControl to a further dimension. Where peak power is so often required only for a limited period, the MultiPlus-II will compensate insufficient generator, shore or grid power with power from the battery. When the load reduces, the spare power is used to recharge the battery. Solar energy: AC power available even during a grid failure The MultiPlus-II can be used in off grid as well as grid connected PV and other alternative energy systems. It is compatible with both solar charger controllers and grid-tie inverters. Two AC Outputs The main output has no break functionality. The MultiPlus-II takes over the supply to the connected loads in the event of a grid failure or when shore/generator power is disconnected. This happens so fast (less than 20 milliseconds) that computers and other electronic equipment will continue to operate without disruption. The second output is live only when AC is available on the input of the MultiPlus-II. Loads that should not discharge the battery, like a water heater for example, can be connected to this output.Virtually unlimited power thanks to parallel and three phase operation (not available for the 8k and 10k models) Up to 6 Multis can operate in parallel to achieve higher power output. Six 48/5000/70 units, for example, will provide 25 kW / 30 kVA output power with 420 Amps charging capacity. In addition to parallel connection, three units of the same model can be configured for three phase output. But that’s not all: up to 6 sets of three units can be parallel connected for a 75 kW / 90 kVA inverter and more than 1200 Amps charging capacity. On-site system configuring, monitoring and control Settings can be changed in a matter of minutes with VEConfigure software (computer or laptop and MK3-USB interface needed). Several monitoring and control options are available: Cerbo GX, Color Control GX, Venus GX, Octo GX, CANvu GX, laptop, computer, Bluetooth (with the optional VE.Bus Smart dongle), Battery Monitor, Digital Multi Control Panel. Remote configuring and monitoring Install a Cerbo GX or other GX product to connect to the internet. Operational data can be stored and displayed on our VRM (Victron Remote Management) website, free of charge. When connected to the internet, systems can be accessed remotely, and settings can be changed. |PowerControl & PowerAssist |Maximum AC input current |DC Input voltage range |38 - 66v |Output voltage: 230 VAC ± 2% Frequency : 50 Hz ± 0,1% (1) |Cont. output power at 25°C (2) |Cont. output power at 25°C |Cont. output power at 40°C |Cont. output power at 65°C |Max apparent feed-in power |Zero load power |Zero load power in AES mode |Zero load power in Search mode |Input voltage range: 187-265 VAC Input frequency: 45 – 65 Hz |Charge voltage 'float' |Charge voltage 'absorption' |Max. battery charge current (4) |Battery temperature sensor |Compatible battery chemistries |Lead-acid, Lithium, Zinc-Bromine and others (5) |External AC current sensor (optional) |50A or 100A |Programmable relay (6) |A - G |VE.Bus communication port |For parallel (not for 8k, 10k and 15k models) and three phase operation, remote monitoring and system integration |General purpose com. port |Operating temperature range |-40 to +65°C (fan assisted cooling) |Material & Colour |Steel, blue RAL 5012 |230 V AC-connection |Screw terminals 13 mm² (6AWG) |Dimensions H x W x D (mm) |499 x 268 x 141mm - Can be adjusted to 60 Hz - Non-linear load, crest factor 3:1 - For Australia only: These inverters are not approved to the standard AS4777.2 2020 and cannot connect to the grid as part of an inverter energy system in accordance with the requirements of AS/NZS 4777.1. For stand-alone applications (generator only) replace the word “grid” in this datasheet with the word “generator”. - Up to 25°C ambient - Other chemistries are possible as well, providing the charger is configured confirm the battery manufacturer’s specifications. - Programmable relay which can be set for general alarm, DC under voltage or genset start/stop function. AC rating: 230V / 4A, DC rating: 4A up to 35VDC and 1A up to 60VDC - Protection key: - a) output short circuit - b) overload - c) battery voltage too high - d) battery voltage too low - e) temperature too high - f) 230 V AC on inverter output - g) input voltage ripple too high
systems_science