url
stringlengths 13
4.35k
| tag
stringclasses 1
value | text
stringlengths 109
628k
| file_path
stringlengths 109
155
| dump
stringclasses 96
values | file_size_in_byte
int64 112
630k
| line_count
int64 1
3.76k
|
---|---|---|---|---|---|---|
https://www.fi.freelancer.com/job-search/installer-excel-macros/ | code | OMO_Import - Error Handling when item prefix identifier does not exist in the master. OMO_Import - Remove filter before starting import (if filter is on) OMO_Import - Better copy of new columns so that formatting and Header formulas are carried across. OMO_Import - Autorun "creatlinks" macro at completion if new items are added Runsheetgen - Borders on either side of drop instances if po...
We have a project which needs to be packaged using PyInstaller or similar and deploy in multiple Operating systems like Windows, Linux and Mac. The project is a simple Web Application developed using Flask. In Windows the Server should start as a service . The other requirement is the Python program should connect to the RS232 device through Serial communication and get/set the data.
I have two excel files with a lot of macros and formulas and different things and I can't figure out how to merge them into one file. One file has like 20+ sheets and the other file just has one sheet and I want to have them all in one file (you can move the 20+ sheets to the 1 sheet file, or move the 1 sheet to the 20+ sheet file, or vice versa).
...have an online business with all of my client records on a web portal. I need to build a database in Excel of all of these records, which requires copy/paste from the website in to excel. I have timed myself doing this, and without the help of any macros, I am able to complete one client record per minute and I have a total of 615 client records to enter
We are using Ch...server via HTPP POST. 7) Repeat from point 3 until the next day, then repeat from point 1. The scripts also can be run manually, for testing. If possible, the new scripts, or macros, or whatever, should run in a standard Windows machine. And have to be editable, inform about what is happening, and have a verbose mode for debugging.
I have an excel spreadsheet with some VB code macros that prompt the user for data input via popup screens, and output some data in a pdf format that I was looking at getting translated into google sheets. This macros that is being used to track a few different things for projects, and was interested in seeing what sort of ballpark timeframe and cost
I have just bought the Web 2.0 directory plugin. The official link is here: [kirjaudu nähdäksesi URL:n] the plugin; - The developer will not have access to the back end of the site. I will only send the zip with the plugin and, after he finishes, he will send me the installer extension of the plugin.
...quality controls. If you can say yes to the following, then perhaps you can join our merry band! 1. Advanced level of excel ability and familiarity of using macros, v-lookups and templates. This means that you are excellent with excel ! 2. Expert data validation skills and meticulous attention to detail. Ability to find the unusual or odd..... Data bods
Hi, We are a Marketing Agency. We need to build a network of blogs and html websites. Here is a list of the tasks: - Installing wordpress themes - Installing wordpress plugins - Adding logos to websites - Adding content to about us - Creating contact form pages - Updating headers and footers - Updating HTML templates with content - Updating HTML templates menus and side bars - Finding images fo...
...I will require screenshots of forms filled initially until trust is built. It's a simple task, and one which could be automated/streamlined to a large degree if you use macros or keyboard shortcut software to paste the email content. (there are some small edits to personalise the emails for each business) (I use short keys lite (free) to insert
I have some work, in an Excel spreadsheet. Looking for a macros to combine multiple csv files and filter based on set of keywords and remove duplicates based on one column.
I have some work, in an Excel spreadsheet. Looking to combine multiple CSV files and filter based on keywords and remove duplicates based on a column. there are two similar csv file formats with one having extra column
Raise & Shine Blinds is a family-owned window covering business, which expanded its operations and services from California to South Florida. Over the past 15 years, we have served the window covering needs of thousands of residential clients, commercial businesses, property management companies, and schools/universities. We are looking for highly experienced and professional window coveri...
...1: [kirjaudu nähdäksesi URL:n] General Overview ===== 2: [kirjaudu nähdäksesi URL:n] Need to remove Touch Screen Trivia logo from installer. Need to replace desktop icon. ===== 3: [kirjaudu nähdäksesi URL:n] There seems to be a problem entering a team name on the sign in screen. The cursor
in the Financial model as enclosed,I am not able to write a macro in this particular file, it says Compile Error. Syntax error. However I can write macros in a different excel file..not sure what is the problem in this file... Can you please help?? [removed by freelancer.com admin] Thanks
we have redirect issue after installing php script [kirjaudu nähdäksesi URL:n] after install the /installer/install/[kirjaudu nähdäksesi URL:n] files keeps opening where as it should be /installer/[kirjaudu nähdäksesi URL:n]
I have an excel workbook that contains a number of spreadsheets. The spreadsheets contain various data - one sheet is a person's roster, another is an extract of their timesheet entries, another is employee data rules. I'm looking for a way to automate the reconciliation process for each employee (over 100) to check if the timesheet entries match the
Looking at creating and automated excel file with VBA and macros for a certain task which would ultimately be automated, save to PDF and .Doc and also enter into another database which will be automatically formed. More details to follow
I have an application that automatically renames movie files and moves them to the correct folder. I need this application to run automaticall...the background whenever there is a change to a specific folder. There are a couple of buttons that need to be clicked automatically as well. I believe I could use a script or macros that can handle this task.
NOTE: Must have experience with coding macros in the Excel PERSONAL file. If you don't know what this means then please do not bid. I have an existing Macro. The function of the macro is to read a sales data file and insert the sales quantities into cells of MASTER file. Task: Extract an Order Number from the data file, and paste into into Row
...applications as it should normally work. coder must submit both Mac and PC applications for me to test and improve throughout the process. at the end you must submit proper installer files For Mac and PC with 15 days trial with Logo built in and 2 installers without logos built in. along with all the code and instructional Video on how to Build the applications
I have the files on a system, I need someone to create an installer with serial! * License Generator * Demonstration license generator I already have the files System is simple Tenho os arquivos de um sistema , preciso que alguém crie um instalador com serial ! *Gerador de licença * Gerador de licença demonstração já tenho os arquivos Sistema
...know what you are doing. Bid with your prior work experience on this. 2. I'll provide Linux VM details (IP address, login name, and password). I'll provide SAP application installer as well (.RAR files). Work instructions are here - [kirjaudu nähdäksesi URL:n] 3. You
Have many multiple excel files. Need macros that will take each main file and split into 100 items per file and save the file as csv Each file should have header labeled SKU and then underneath it should have 100 sku's. Each file should be saved as CSV. It should only take the items in the first column... Should be done in 24 hrs...
A store's products are represented in a excel file. Each sheet in the file represents a shelf layout of a fixture, each column is the row of the fixture from left to right. The idea is to write a program that takes a Planogram file as a template and create a graphical layout by replacing the UPC numbers with the actual product images. The images are
Step 1: Need to remove "continue" button from checkout page and allow re...checkout page and allow redirect to the payment gateway after filling the checkout form instead of showing the information. (please check the screenshots) Step 2: Create macros of the products for custom emails. Step 3: Clear cart/session whenever we close and an open tab.
...17.9 which can be installed by an administrator but run by an users. Use this code as a template, however it does not run for all users. It needs to be wrapped into an installer. [kirjaudu nähdäksesi URL:n] A similar product which works for all users is redgate search https://www
Need Macros expert to make excel based tool for Lead Management and Delivery Management system. entries would be created with help of forms only. emails to be sent to dynamic list of receipients based on schedule or manual trigger. Data to be copied to other excel using macros.
...installers in one click, as opposed to the tedious task of building the app, finding installer creation software, writing more code etc etc. It's called "Merlin - Installer & Build Automation". It's named after Merlin the wizard, taking a twist on the idea of an installer being called a "Setup Wizard". So I thought the name Merlin was quit...
...windows machines and willing to install my software for testing purpose, my company is developing a new software and all i need from you is: 1) Download the .exe file or installer 2) Install o run the .exe or install the app and send me back the logs. My budget is USD 25 for this task, I need windows machines with different OS versions, could be XP | s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039742253.21/warc/CC-MAIN-20181114170648-20181114192648-00139.warc.gz | CC-MAIN-2018-47 | 9,803 | 30 |
https://www.zdnet.com/article/microsoft-crm-whats-coming-in-the-next-release/ | code | Microsoft has posted a preview as to what's going to be in the next version of its Dynamics CRM product.
The cloud version -- CRM Online -- is going to be first to get a number of new updates, sometime in the coming weeks. Microsoft will make those same updates, plus some additional features, available via an on-premises version of its CRM release. I believe (but am not totally sure) both of these updates are coming in Q4, based on a new post on the Dynamics CRM blog.
Microsoft execs said earlier this year that the CRM team is moving to a twice-yearly delivery schedule for both the CRM Online and CRM on-premises offerings.
From the October 19 CRM blog post:
"Microsoft Dynamics CRM is moving to a rapid innovation cadence where additional capabilities for Microsoft Dynamics CRM (both Online and on-premises) are delivered in releases targeted for the Spring (Q2) and Fall (Q4) of each calendar year. The first of these releases will be the Q4 2011 Service Update."
(The wording makes me uncertain whether the "Service Update" is a CRM Online only thing and the next on-premises update will be in spring 2012.)
New "activity feeds" -- which are based on Facebook and/or Salesforce Chatter's model (take your pick) are coming to both the online and on-premises versions of the next Dynamics CRM release, the blog post notes. Microsoft also is going to offer a Windows Phone version an activity feeds client as part of this update. (This is not a full CRM client for Windows Phone; it is just a feeds client.) New business-intelligence capabilities and enhanced data-cleansing functionality are also coming to both the online and on-premises releases, Microsoft officials blogged. | s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474715.58/warc/CC-MAIN-20240228112121-20240228142121-00005.warc.gz | CC-MAIN-2024-10 | 1,686 | 7 |
http://technomixer.blogspot.com/2011/01/download-for-free-microsoft-onenote-app.html | code | Shares a similar feature set to its cousin on Windows Phone 7, OneNote Mobile for iOS lets you create and view notes and lists whenever you need them. Notes are automatically backed up and synced with free Windows Live SkyDrive online storage, so that you can access them from virtually anywhere using your phone, PC, or almost any web browser. Unfortunately, The new iOS mobile app isn’t as capable as the Web version or the Windows desktop client as it doesn’t supports tables, hyperlinks, uploaded graphics,clip art, and some of those advanced data types and formats.
With OneNote Mobile for iOS, you can:
- Create flexible notes that can include text, pictures, bullets, and checkboxes
- Check To Do items off on the go
- Save time with quick access to your most recent notes
- Work with confidence—OneNote Mobile automatically syncs your notes with Windows Live in the background
- Organize your notes into sections or create new notebooks using OneNote 2010 or OneNote Web App and access them from your iPhone
It will work on any iOS device that is running iOS 4.2 or newer though it’s not optimized for the iPad’s larger screen, and requires a Windows Live ID. You may also like to read A first look at Microsoft OneNote on the iPhone to see how to install, start, use, and sync the new app. | s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463607786.59/warc/CC-MAIN-20170524035700-20170524055700-00246.warc.gz | CC-MAIN-2017-22 | 1,308 | 8 |
http://tkost.tk/name/isp-programming/ | code | Target Socket Section
In-System Programming uses the Atmel AVR internal SPI (Serial Peripheral Interface) to download code into the flash and EEPROM memory of the AVR. ISP programming requires only VCC , GND, RESET and 3 signal lines for programming. All AVR devices except AT90C8534, Attiny11 and ATtiny28 can be ISP programmed. The AVR can be programmed at the normal operating voltage, normally 2.7V-6.0V. No high voltage signals are required. The ISP programmer can program both the internal flash and EEPROM. It also programs fuse bits for selecting clock options, startup time and internal Brown Out Detector (BOD) for most devices.
Note that the ISP frequency (SCK) must be less than 1/4 of the target clock. However this requires a 50/50 dutycycle on both target clock and ISP clock. Running a ISP frequency at 1/5 or less than target clock is recommended on Atmel STK500. The ISP frequency is set by the board tab of the STK500 user interface.
High-Voltage programming can also program devices that are not supported by ISP programming. Some devices require High-Voltage programming for programming certain fuse bits. See the High-Voltage programming section for instruction how to use High-Voltage programming.
Because the programming interface is placed on different pins from part to part, 3 programming headers are used to route the programming signals to the correct pins. A 6-wire cable is supplied for connecting the ISP signals to the target ISP header. A color coding system, and a number system, is used to explain which target ISP header is used for each socket.
During ISP programming the 6-wire cable must always be connected to the header marked ISP6PIN. When programming parts in the blue sockets, connect the other end of the cable to the blue SPROG1 target ISP header. When programming parts in the green socket, use the green SPROG2 target ISP header. And when programming parts in the red sockets, use the red SPROG3 target ISP header. The table below shows which socket suits which AVR device, and which SPROG target ISP header to use for ISP programming.
The 6-wire cables should be connected directly from the ISP6PIN header to the correct SPROG target ISP header. The cable should not be twisted. A colored wire on the cable indicates pin one. Confirm that this is connected to pin one on each of the headers.
When programming 8-pin devices, note the following: Pin 1 is used both as RESET and as PB5 on some devices (ATtiny11, ATtiny12, ATtiny13, ATtiny15 and ATtiny45). Pin 1 on the 8-pin sockets SCKT3400D1 and SCKT3600A1 are connected to PB5. The RESET signal used during ISP programming is therefore not connected to pin 1 on these sockets. This signal must be connected by placing a wire between RST on the PORTE header and PB5 on the PORTB header. Use one 2-wire cable to connect the PB3 pin (pin 4) on the PORTB header to the XT1 pin (pin 7) on the PORTE/AUX header. This will connect the clock system to the AVR device.
Connect RST on PORTE to PB5 on PORTB.
Connect XT1 on PORTE to PB3(XTAL1 on 2323) on PORTB.
SPROG3. These are MCS51 core compatible parts, they do not contain an AVR core. Only ISP programming is supported on these devices, no other functionality.
Connect RST on PORTE to PB5 on PORTB.
[a] This device is also supported on the STK505 Top Module
The figure below shows an example of how ATmega8515 can be In-System Programmed. The 6-wire cable is connected from the ISP6PIN header to the red SPROG3 target ISP header, and the ATmega8515 part is inserted in the red socket marked SCKT3000D3.
It is not necessary to remove the 6-wire cable from it’s ISP position while running a program in the AVR. The port pins used for ISP programming can be used for other purposes in your program. | s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267155814.1/warc/CC-MAIN-20180919004724-20180919024724-00207.warc.gz | CC-MAIN-2018-39 | 3,744 | 15 |
http://blogs.msdn.com/b/ukhe/archive/2013/11/22/friday-recap-post.aspx | code | Check out the links below for a recap of all of this weeks blog posts.
Have a great weekend.
DreamSpark Student Review: Alex Furnell, University Student
Are Tablets the biggest transformation in education since the internet?
UTC Reading students are first in the UK to get professional Microsoft qualifications and become IT Academy
Flipping the Classroom
Importance of Computer Science in the Curriculum Conference, sponsored by ASCL and Microsoft | s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701165302.57/warc/CC-MAIN-20160205193925-00080-ip-10-236-182-209.ec2.internal.warc.gz | CC-MAIN-2016-07 | 448 | 7 |
https://www.ukpunting.com/index.php?topic=45445.125 | code | I think your view would be genuinely welcomed MrTrelf - despite what has gone before on this thread, in my mind, your view will always be relevant
My views have not really changed as I have no proof one way or the other as to what is true.
All I can say is what I already have said....which is that in my experience there hasnt been any time when I have been on a booking with Amour and there has been more than one escort in the flat.
Also that all my dealings with that agency have been straight forward and honest.
I am nothing to do with Stacey or anyone at Amour,other than a satisfied customer.
I know nothing about threats of any kind to anyone so have not made comment.
If proven then I would obviously condemn it. | s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917122720.81/warc/CC-MAIN-20170423031202-00017-ip-10-145-167-34.ec2.internal.warc.gz | CC-MAIN-2017-17 | 722 | 7 |
https://www.pinterest.com/pin/98023729361185455/ | code | 20+ Wire Jewelry Making Ideas Here’s a nice mix of jewelry making tutorials using wire, beads, seaglass and gemstones. A little bit of everything in this bunch, I’ll be adding more to this project sheet as I find them. Don’t miss the resources I added at the bottom where there are dozens more tutorials available. Enjoy!
The "EZ Bracelet" bracelet sizer is a must for every beader. Make bracelets that fit the first time, every time! No more re-stringing because it's the wrong size! No more wasted beading wire! No more wasted crimps! No more wasted time! | s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463608107.28/warc/CC-MAIN-20170525155936-20170525175936-00004.warc.gz | CC-MAIN-2017-22 | 563 | 2 |
https://www.controleng.com/articles/benefits-of-learning-ladder-logic-for-industrial-programming/ | code | Benefits of learning ladder logic for industrial programming
Ladder logic is challenging for industrial programming, but it is a valuable skill for engineers looking to enhance their skill set.
- Ladder logic is a challenging programming language to learn for those not familiar with industrial programming.
- Users familiar with object-oriented language (OOL) may not be used to the more manual nature of ladder logic.
- Ladder logic is a product of an industry that has been slow to adapt to the more modern approaches.
I started programming as an early teen, received a bachelor’s degree in computer science and I’ve worked as a software and web developer for over a decade. I’ve been a programmer before attaining the degree: teaching myself, learning new languages and trying new things. Transitioning to industrial programming was the biggest changeup in my history, but also a valuable one.
Adding ladder logic into industrial programming
The first thing when adding ladder logic to industrial programming is it looks like a bad visual integrated development environment (IDE). Like somebody tried to give you a flowchart you could drag and drop onto. Considering it’s mostly a mix of logical gates, it really does behave like that.
Once you start to tinker, get your head around it, it’s usable. Basic structures like loops are a mess, sure, and the variable structure is ugly. The whole program is stuck in an infinite loop. The thing that’s going to slow you down the most is how much you need a mouse.
With object-oriented language (OOL) such as Java, VB, or any common scripting language, navigation requires very infrequent mouse usage. Get to the place you want to edit and start typing. There’s a double-click here and there, but it’s mostly typing because it’s all characters and symbols. To speed things up, many IDEs have an auto-completion function built-in.
Navigation inside PLC programming
In ladder logic, keyboard shortcuts can keep things moving, but navigating between rungs with only the arrow keys can be tedious, slowing down the execution, and derailing your train of thought. Ever walk into a room and forget why you went in there? Imagine that happening every 10 seconds.
OOL and a modern IDE provide autocompletion of variables and method names, easy navigating, better code organization and structure and more effective use of screen real estate. Instead of a bunch of boxes and arrows to space things out, all that appears are letters and numbers. The user can see much more information without having to scroll around.
The biggest hurdle when switching to industrial programming is how slow it is. It’s a beast of its own in a variety of ways, and there’s only so much that can improve speed. Things are going to take longer to write in ladder. It’s a given.
Looking forward: PLC programming needs more PC programmers
As for the other issues, they’re remnants of the early days of an industry that has been slow to adapt. That is exactly why I think the industry needs more PC programmers. The manufacturing industry hasn’t seen significant changes to its programming methodologies in decades despite immense improvements in other fields. It needs fresh ideas and modern concepts.
PLC programming is not fun, not efficient and not easy to maintain. And yet, it remains a very worthwhile skill. Frankly, the more PC people who learn it, the faster we can help the industry evolve its methods to more accessible and maintainable solutions.
David Breen, lead programmer, Breen Machine Automation Services, a CFE Media content partner. This article originally appeared on Breen Machine Automation Services’ blog. Edited by Chris Vavra, associate editor, Control Engineering, CFE Media, [email protected].
Keywords: PLC programming, ladder logic, object-oriented language
Ladder logic is a challenging programming language to learn for those not familiar with industrial programming.
Users familiar with object-oriented language (OOL) may not be used to the more manual nature of ladder logic.
Ladder logic is a product of an industry that has been slow to adapt to the more modern approaches.
Read this article online at www.controleng.com for additional stories about ladder logic.
What benefits did you gain from learning ladder logic? | s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474361.75/warc/CC-MAIN-20240223053503-20240223083503-00244.warc.gz | CC-MAIN-2024-10 | 4,312 | 24 |
http://www.coderanch.com/t/345964/GUI/java/logic-thread | code | I am trying to add/store these sets of values into a hashmap. I store the labelName(String) and capture corresponding checkbox valuess(integer array) into an integer array. Then I add labelName as key and check box values as value into hashmap. My logic works as below:
My logic is, I am checking every component. If it is a label, I am assiging it to "key" value. As long as the next coponent of check box is not label, I am collecting it into int array. (Basically I am scanning through every component in the panel) So if teh next element is label, the time to add key and values into hashmap and will start scanning next line.
You are trying to increment the component index, and no component exists. That is causing the problem. We had discussed this in your previous thread. Since the pattern is label, checkbox, label checkbox... you dont really need to increment the counter because you are always going to encounter the components in the above mentioned order.
Another way (though unnecessarily complicated) way of doing the same this is create a custom panel, which has label and checkbox embedded inside it. You add these panels to your base container. Thus, when you iterate through the base container, all the components you will get are your custom panel. Then you can invoke the appropriate methods on the custom panel to get the required values.
yes, I got it. That makes sense. B'cos My panel would contain, one label followed by any number of check boxes. At the end of check box list, I need to add these key and values into hashmap. And then start scanning next line(ie label). This is what I am trying to do. I will try to add the components or group them into panels and try to add it into the final dPanel. Thanks.
First of all, String comparison should use equals: name.equals("javax.swing.JCheckBox"). Otherwise you are checking for the same reference, and that check fails in most cases.
Second, you don't need to check for the name - check for the Class object itself:
Now you notice that here I do use reference checking. This is because (unless you're using different ClassLoaders and forcing a ClassLoader to load its own version of a class - something not possible for the classes from the Java API) there is just one Class object for each Class, and it can be references through <Class Name>.class as well as through <instance of class>.getClass().
Hi, still I have'nt found a solution for this problem. Is there anyway to collect the label name in the key field and selected checkboxes as arrayList/list in the value field of a hashmap. Thanks.
Joined: Jan 13, 2008
I tried adding every set of label and checkboxes in one panel and then adding into final panel. But when I try to add components, it is not able to count all internal panels inside the main panel Hence I went back to my original logic as below:
It is serving the purpose and printing label followed by selected check box. Now I need to add label into key value and all the selected checkbox into value followed by next set of label and check boxes. Kindly suggest. Thanks. | s3://commoncrawl/crawl-data/CC-MAIN-2015-48/segments/1448398446250.47/warc/CC-MAIN-20151124205406-00151-ip-10-71-132-137.ec2.internal.warc.gz | CC-MAIN-2015-48 | 3,071 | 12 |
https://openlmis.atlassian.net/browse/OLMIS-5387 | code | Currently, all commits pushed to a dev branch (non-master and non "rel-" prefixed) trigger a build. This is fine since we want the developers to get instant CI feedback on their changes, but they should not notify the Slack #build channel on failures.
CI continues to build changes pushed to a dev branch
Slack's #build channel does not report if the pipeline execution fails on the master branch
Slack's #build channel still reports if pipeline executions fails on master or "rel-" branch | s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439736962.52/warc/CC-MAIN-20200806121241-20200806151241-00570.warc.gz | CC-MAIN-2020-34 | 489 | 4 |
http://archives.miloush.net/michkap/archive/2008/09/13/8950706.020100.html | code | by Michael S. Kaplan, published on 2008/09/14 02:01 +00:00, original URI: http://blogs.msdn.com/michkap/archive/2008/09/13/8950706.aspx
Regular readers (and devoted couple!) Don and Tammy asked via the Suggestion Box:
We found another interesting language issue in adult films. This time it is in the title.
There is a long-running series from Vivid entitled _Where_the_Boys_Aren't_. As you can probably guess if you have never seen any of them, it is an "all girls" series. It started some time in the early '90s and new ones are still coming out to this day. Some are good, some not. Just like with mainstream movies.
But the point is that there are 19 different releases. The meaning of "Where the Boys Aren't 16" and "Where the Boys Aren't 17" and "Where the Boys Aren't 18" and so forth seems like a confusing use of, doesn't it? We were joking about this while watching the latest one, and decided to forward it to you since I think you haven't covered it before.
Don (and Tammy)
Well, as Don allowed for, I was unaware of the specific series. My experience is these things is much more limited then theirs, but what is the sense of reader feedback if not to afford me the chance to live vicariously through them? :-)
Of course if the goal is to allow for the confusion (you know -- to have fun with it) then perhaps it is intentional, though if every one of them since the beginning had a number on it, I doubt that it started that way -- no one is really expecting anything sordid out of a title like Where the Boys Aren't 3, for example. Plus with lower numbers it is much more common to use the spelled out number which would be even less confusing, I think? Again, I'm kind of guessing on some of this.
Though this perhaps kind of a garden path sentence, like in this blog or this one or this one or this one or this one or even this one. Though in this case for many it is not so much a garden path at all -- it is just a regular old path that one may choose to be walking down if one likes. Either one successfully parses it and thinks of
Where the Boys Aren't 20
as if it were
Where the Boys Aren't, #20
and not, as if it were
Where the Boys Aren't 20 years old
depending (I suppose) on one's frame of mind or movo\ie preference.
The two paths here suggest two completely different types of movies with two completely different target markets -- each of which might contain members very willing to skip the movie if they read the sentence the other way, which would suggest that Vivid would be better off putting the # in the title since the "pun" is dumb and the misunderstanding doesn't really help them make extra sales (in fact, customers expecting a "no guys" feature are almost certainly going to run away from a "guys of a specific age" feature!).
Though I doubt that the company producing the series has on-staff linguists to control title quality, and not only am I not a linguist but even if it were okay under the moonlighting clause it would probably fail for "moral turpitude" reasons. So there is no sense trying to ask for the job....
With a series running for so long that (assuming 1992 to 2008 with #18 as the last one, a little over one a year) people might really know the series well enough to have no confusion, it may never come up as an issue in practice.
I know that it might be awkward to go the video counter to complain about being misled, at least....
Either way, it is sort of a garden path sentence kind of thing I guess, though the path may not be heading toward the most appropriate destination for all of us (as is the title of this blog, only moreso)! :-)
This blog brought to you by ⚢ (U+26a2, aka DOUBLED FEMALE SIGN)
go to newer or older post, or back to index or month or day | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100575.30/warc/CC-MAIN-20231206000253-20231206030253-00553.warc.gz | CC-MAIN-2023-50 | 3,730 | 22 |
https://datisdesign.com/portfolio-item/sync-slicers/ | code | One of the problems with multiple pages report before releasing this feature was that users were not able to transfer their selected filters on a specific page to other pages in their report. They had to create same filters on each page and select same filter each time they navigate through different pages.
This feature helped them to solve this issue by creating filters that are synced in entire pages of their report and each they select a filter in one page, it carries through all other pages without doing same action over and over again.
We designed 3 different concept for this scenario and tested them with users.
Global Slicer Switch in Property Pane
Advanced Slicer Management Pane
Simple Slicer Management Pane
One Of The Top Requests From Power BI Community
In this concept we tried to put controls of sync slicers inside property pane of visual.
3. User needs to create a group name for this new sync slicer
4. This option allows them to show this slicer on all pages of report
After creating a new group, adding more groups inside the group selection drop down was challenging and ended up adding more UX problems.
Since any change that made on property is live, having an input box without save button changes the pattern in property pane.
In this concept we tried to put controls of sync slicers inside a sidebar pane.
3. After saving group name, they see their report structure that shows the slicers they have in each of the pages of their report. By clicking on sync icon they can sync any slicer in any page with selected slicer (if they are compatible). They can control the visibility of slicers by clicking on eye icon
Icons for sync and visibility are not intuitive and not clear the what are the related actions.
4. If user selects another page, canvas switches to that page to help user select the proper slicer from the list.
5. Also if any of the slicers in the page are incompatible with selected slicer, the error icon appear on its row to notify user about incompatibility
Showing slicers that are incompatible with user selected slicer in this pane would add to user confusion.
Since users preferred the pane design we decided to iterate on it and make it simpler based on user feedback.
4. From here user can select the page from table or by clicking on pagination on canvas and add/hide slicers to pages. (By clicking on page row in table, report page changes to that selected page). Title will change based on selected slicer.
After selecting a slicer almost all users were able to sync the slicers with other pages in their report. | s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224652184.68/warc/CC-MAIN-20230605221713-20230606011713-00196.warc.gz | CC-MAIN-2023-23 | 2,570 | 21 |
https://documentation.maptiler.com/hc/en-us/articles/4410767180817-MapTiler-Server-user-interface | code | This article describes the user interface of the MapTiler Server. After you successfully install
MapTiler Server on Linux, Windows, or Mac, your server should (by default) be available on the address http://localhost:3650. If you open it in a browser, you will see a blank screen with a MapTiler Server logo.
Signing in to the administration
The administration is available on http://localhost:3650/admin. Fill in the password you set during the installation, or which was generated automatically (and which is visible in the terminal window).
After you sign in, you will see the landing page with your maps. It is going to be probably blank at first.
The second area is called Tiles. Paste your geodata (in MBTiles or tiled GeoPackage format) into a predefined folder, click on the Re-Scan button and all datasets will be visible in the administration.
Click on the Publish button for each dataset you want to be published.
Click on the Details button if you want to take a closer look at an individual data set.
On the new page, you can manipulate the dataset, see details, and get links for map services.
You can open an Embeddable viewer in the new window and browse your dataset.
Web map service (WMTS) GetCapabilities doc. is available on the page with the dataset and URL /WMTS.
The whole setting section is divided into five parts:
In the Server section, you can set up the behavior of the MapTiler Server.
On the Security page, you can change your password.
The hash of your password is stored in the config.json file: if you delete it, a new password will be automatically generated for you.
Another option to change your password is to run it from a command line by appending the --adminPassword suffix.
Rasterization is fully supported on Windows and Linux. Mac users requiring rasterization can use Docker image.
Unlock the full potential of your MapTiler Server with a valid License Key.
This section shows information about the software you are running (version, build date, ...). This is very important when you are submitting a bug report. | s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474526.76/warc/CC-MAIN-20240224080616-20240224110616-00244.warc.gz | CC-MAIN-2024-10 | 2,056 | 19 |
https://se.mathworks.com/help/reinforcement-learning/ug/specify-simulation-options-in-reinforcement-learning-designer.html | code | Specify Simulation Options in Reinforcement Learning Designer
To configure the simulation of an agent in the Reinforcement Learning Designer app, specify simulation options on the Simulate tab.
Specify Basic Options
On the Simulate tab, you can specify the following basic simulation options.
|Number of Episodes||Number of episodes to simulate the agent, specified as a positive integer. At the start of each simulation episode, the app resets the environment.|
|Max Episode Length||Number of steps to run the simulation, specified as a positive integer. In general, you define episode termination conditions in the environment. This value is the maximum number of steps to run in the simulation if those termination conditions are not met.|
|Stop on Error||Select this option to stop simulation when an error occurs during an episode.|
Specify Parallel Simulation Options
To simulate your agent using parallel computing, on the Simulate tab, click . Simulating agents using parallel computing requires Parallel Computing Toolbox™ software. For more information, see Train Agents Using Parallel Computing and GPUs.
To specify options for parallel simulation, select Use Parallel > Parallel training options.
In the Parallel Simulation Options dialog box, you can specify the following training options.
|Transfer workspace variables to workers|
Select this option to send model and workspace variables to parallel workers. When you select this option, the parallel pool client (the process that starts the training) sends variables used in models and defined in the MATLAB® workspace to the workers.
|Random seed for workers|
Randomizer initialization for workers, specified as one of the following values.
|Files to attach to parallel pool||Additional files to attach to the parallel pool. Specify names of files in the current working directory, with one name on each line.|
|Worker setup function||Function to run before simulation starts, specified as the name of a function having no input arguments. This function is run once per worker before simulation begins. Write this function to perform any processing that you need prior to training.|
|Worker cleanup function||Function to run after simulation ends, specified as the name of a function having no input arguments. You can write this function to clean up the workspace or perform other processing after training terminates.|
The following figure shows an example parallel training configuration the following files and functions.
Data file attached to the parallel pool —
Worker setup function —
Worker cleanup function —
- Design and Train Agent Using Reinforcement Learning Designer
- Specify Training Options in Reinforcement Learning Designer | s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224652235.2/warc/CC-MAIN-20230606045924-20230606075924-00418.warc.gz | CC-MAIN-2023-23 | 2,718 | 24 |
https://www.webmasterworld.com/google/4177761.htm | code | Well it seems kind of an obvious thing to test, everyone is accepting that nofollows dont pass pagerank and hence a site cant rank using them alone.
But I believe google still uses predominantly incoming links to rank a site and since every man and his dog is merrily nofollowing every link google's algorithm would be broken by now, so I think there is a lot of FUD spread by them and they still have a benefit.
Google said nofollow does not passes PR, they did not say that nofollow links are ignored in ranking alghorithm - which, as we know, has more than 200 factors. I am also really interested in your experiment as my opinion of nofollows is the same as yours. I also think citations help too, perhaps indirectly via trust factor.
If you imagine a really authoritative and trusted site giving you a link - great. The same site giving you nofollow link - I think it would still count somewhere. Not giving you link, but there is a "citation" - expecially if the text of the page or aroud citation is semantically connected to your niche/keywords... I think this helps too as it shows some kind of "popularity", albeit not measured via PR, but there is still some kind of "web popularity".
"since every man and his dog is merrily nofollowing..."
It could reach an overdone level and my feeling is Gorg didn't realize it could cause a problem. Perhaps small for now, but how about in a few years? I noticed a smallish site I exchanged links with five years ago(same industry) whom I never thought knew much about SEO started, no following. Thanks for telling me fella...his link was yanked.
Too cynical for my taste, anyway, yes. I've even been considering a separate thread about how both the Google-fan and the Google-critic points of view can cloud judgment.
Besides, I don;t think the really problematic spammers would stop anything they are doing unless it stopped working - doesn't matter what Google says. So the anit-spam workload would not fall off in the least.
I have at least one page with a tool bar PR that cannot be explained by its incoming links. It does not have many internal links, and the only significant external links are multiple links from Wikipedia (all no-followed).
The other explanation is that toolbar PR is not just incomplete or out of date, but a flat out lie. | s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496665976.26/warc/CC-MAIN-20191113012959-20191113040959-00315.warc.gz | CC-MAIN-2019-47 | 2,299 | 10 |
https://wilkiecat.wordpress.com/2014/01/01/how-do-i-restore-my-pis-sd-card/ | code | You’ve been playing with the Pi, and you’ve messed up the operating system, so how do you go about making a brand new Pi?
Take the SD card out of your Pi and put it into your dual boot Ubuntu laptop. On OS Elementary (or any other Distro) use the Gparted tool to delete all the corrupt partitions. This tool is called “Disk Utility” on OS Elementary, it is in the Applications folder.
On the Windows side of the machine, you can use EASUS partition manager to do this, but this program is slow.
If it was originally an 8gig SD card, it will have been subdivided into smaller partitions by the Pi’s auto-setup procedure. You need to make sure that all the partitions are deleted, and that the whole SD card is re-formatted as FAT32.
You can do the re-format on the Windows side of the machine using the “SD formatter” program, which is recommended by the Raspberry Pi website. Once the SD card is clean, you write a disk image of Raspian to it using “Win32DiskImager.exe”, the recommended program.
Caveat: re-formatting of SD cards is or is not supported on Linux, depending on your hardware and distro. That’s why we have a dual boot set up!
The disk image you want to download is at this URL
However, the web page advises you to download
New Out Of Box Software
We advise you not to take this option because the download is over 1gigabyte.
Instead, go down to:
Are the other operating Systems any good?
Well that’s the $64 000 dollar question.
Find out. Get another 8gig SD card, download the NOBBS Zip file, and expand it onto the card. You can do this on LInux. Just select the SD card as the destination folder.
The program you want for doing the download is called “uGet” which is in the Software Centre. Copy and paste the URL, to uGet which is currently:
Copying and pasting from here will work, if this post is not out of date. | s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084889917.49/warc/CC-MAIN-20180121021136-20180121041136-00453.warc.gz | CC-MAIN-2018-05 | 1,861 | 16 |
https://sourceforge.net/directory/language:shell/language:jsp/os:os_portable/ | code | With Infoblox DDI, your IT teams can consolidate DNS, DHCP, IP address management, & other core network services into a single platform. Orchestrate DDI functions for hybrid & public cloud & virtual & private cloud environments and access rich reporting & analytics capabilities for capacity planning, asset management, compliance control, & auditing. Boost IT efficiency & automation with seamless RESTful API integration.Sponsored Listing
- Audio & Video
- Business & Enterprise
- Home & Education
- Science & Engineering
- Security & Utilities
- System Administration
Solución de Gestión Empresarial en Software Libre7 weekly downloads
Jdptool is a debug and profiler tool which based on JPDA; It provide both swing based and browser based GUI. It has sophisticated debugger without user interfere. It also provide profiler to memory usage and threads with numerous new feature
The book shows you how architecting a modern security visibility infrastructure can help your business adapt to new threats, lengthen the usefulness of your current security investments, and optimize uptime and availability. See the entire threat landscape, understand security visibility basics, support both inline and out-of-band security systems, explore security visibility use cases and ask the right questions about visibility.Sponsored Listing
KwanPlate is a collaboration-oriented content management system that facilitates integration with relevant existing resources.
SCHOOLDATA e' un software per la gestione dei documenti didattici di una scuola. | s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917121267.21/warc/CC-MAIN-20170423031201-00277-ip-10-145-167-34.ec2.internal.warc.gz | CC-MAIN-2017-17 | 1,542 | 12 |
https://www.developrec.net/job-tech/contract-product-manager-bbbh4433-1685528909 | code | Outside IR35 Product Manager £400 per day 1 Month Discovery Contract
We are partnered with a Digital Agency who are working with one of the UK's biggest beverage suppliers to overhaul their online presence including building a new eCommerce platform.
This will be an initial 1-month discovery contract in which you the Product Manager will flesh out the project and what will be required to complete the project on time and on budget. It will then turn into a 6-month rolling contract should all go well.
You will need to have the following skills and experience:
Nice to haves:
£400 per day
Remote with maybe the occasional need to travel to London for a meeting but the client is flexible on this. | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100762.64/warc/CC-MAIN-20231208144732-20231208174732-00858.warc.gz | CC-MAIN-2023-50 | 701 | 7 |
https://www.freelancer.com.ru/projects/php/fix-sql-bug-whmcs-croster/ | code | On my frontpage footer, I have an entry "blog", this entry does not exist in my admin panel, that means there is a bug in sql database and you need to find and fix the wrong entry in footer menu.
11 фрилансеров(-а) готовы выполнить эту работу в среднем за $25
Hi, I'd like you to send me a link to your website, for me to take a look at the footer and see what can be wrong. Maybe it's not about the database, but about some of the code files. Thanks!
I am a Oracle Developer and also good knowledge of SQL, MySQL. Also i am working on two years a IT Company. That's reason I have on SQL and MySQL experience. | s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656104669950.91/warc/CC-MAIN-20220706090857-20220706120857-00387.warc.gz | CC-MAIN-2022-27 | 655 | 4 |
https://www.pdl.cmu.edu/PDL-FTP/PASIS/Ganger01_abs.shtml | code | Survivable Storage Systems
DARPA Information Survivability Conference and Exposition (Anaheim, CA, 12-14 June 2001), pages 184-195 vol 2. IEEE, 2001.
Gregory R. Ganger, Pradeep K. Khosla, Mehmet Bakkaloglu, Michael W. Bigrigg, Garth R. Goodson, Semih Oguz, Vijay Pandurangan, Craig A. N. Soules, John D. Strunk, Jay J. Wylie
Electrical and Computer Engineering
Carnegie Mellon University
Pittsburgh, PA 15213
Survivable storage systems must maintain data and access to it in the face of malicious and accidental problems with storage servers, interconnection networks, client systems, and user accounts. These four component types can be grouped into two classes: server-side problems and client-side problems. The PASIS architecture addresses server-side problems, including the connections to those servers, by encoding data with threshold schemes and distributing trust amongst sets of storage servers. Self-securing storage addresses client and user account problems by transparently auditing accesses and versioning data within each storage server. Thus, PASIS clients use threshold schemes to protect themselves from compromised servers, and self-securing servers use full access auditing to protect their data from compromised clients. Together, these techniques can provide truly survivable storage systems.
FULL PAPER: pdf | s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964358323.91/warc/CC-MAIN-20211127223710-20211128013710-00066.warc.gz | CC-MAIN-2021-49 | 1,331 | 8 |
https://github.com/silexlabs/Cocktail/tree/c0d2aac0d430666c3bae1b4ae1d0fabbc0fa30ef | code | HTTPS clone URL
Subversion checkout URL
An HTML/CSS rendering engine for the Haxe language
fixes-for-openfl fullscreen-pseudo-class gh-pages layout-refactor master non-interactive-scrollbar refacto refactor-boxmodel refactor-embedded-block-layout update-samples
Nothing to show
Fetching latest commit...
Cannot retrieve the latest commit at this time.
|Failed to load latest commit information.| | s3://commoncrawl/crawl-data/CC-MAIN-2015-48/segments/1448398456289.53/warc/CC-MAIN-20151124205416-00012-ip-10-71-132-137.ec2.internal.warc.gz | CC-MAIN-2015-48 | 395 | 8 |
https://bitsypool.com | code | Register account with your FaucetPay BTC address.
Claim faucet while mining to increase your profits!
Invite your friends and get 50% commission from faucet claims and 10% from mined hashes.
You can use the embedded webminer first but GPU miner is much more efficient.
Download miner below, which will automatically detect your hardware. On first run miner will guide you through setup and start mining.
GPU miner is XMR-STAK, you can alternatively download them from the developers official Github pages. | s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817249.26/warc/CC-MAIN-20240418222029-20240419012029-00242.warc.gz | CC-MAIN-2024-18 | 505 | 6 |
http://wiki.squeak.org/swiki/26 | code | Swiki Color Schemes
This scheme looks somewhat like Windows 3.1. It is a good conservative color scheme.
This scheme was made to be small and unobtrusive. Unlike the others, it shows up to the right of the title in the upper right-hand corner. For those who like their screen real estate, this is it.
This scheme was made to look very simple and text like. It is somewhat inspired by Jakob Nielsen's http://www.useit.com/.
DockAnother one inspired by MacOS. | s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376826306.47/warc/CC-MAIN-20181214184754-20181214210754-00196.warc.gz | CC-MAIN-2018-51 | 457 | 5 |
https://bbpress.org/forums/topic/stopping-spam-and-front-end-sign-up/ | code | Stopping spam and front end sign up?
What is the best way to get users to sign up to your site without letting automated spam bots in and to make things frontendy?
I have searched and searched yet the only option I can find are those premium plugins (which have many vulnerabilities in them).
what I want to to do is simple:
Have a front end registration page with fields for login, password, why does this person want to join… and a custom captcha question… and then somewhere in the admin panel I want to be able to just approve or decline users based on what they put in the why do you want to join field.
I also want to make this the only place for users to register from… else spammers will just use the normal wordpress sign up to get access.
How can I go about doing all of this?
- You must be logged in to reply to this topic. | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100164.87/warc/CC-MAIN-20231130031610-20231130061610-00702.warc.gz | CC-MAIN-2023-50 | 840 | 8 |
http://en.kafra.ru/t225521.html | code | - Added new quest window entries for the remaining 2nd job quest skills (Assassin/Wizard/Sage/Dancer/Rogue).
- Corrected the Eyeball Ring's option of CRIT+2, MaxSP+25.
- Fixed the small water jug's options "for 90 seconds your weapon becomes water element".
- Corrected the Dead Branch Staff's option "Each refine above +6 gives 1 int, MaxHP - 200, MaxSP -100", and corrected the number of card slots shown.
- Changed the following MATK display for the staffs from % MATK to +MATK. (Lacrima Stick, Croce Staff, Chronos, Staff of Thea, Staff of Orde)
- Corrected the Tae Goo Lyeon 's option not applying for 3rd job characters at job level 50.
- The Rank Prize Event reward NPC was added.
> The rewards can be received from the Jonda NPC in town.
> An event NPC is added to the prontera blacksmith guild.
> The item "Refine Guarantee Ticket" can only be used at this NPC.
> For more details for the Refine Guarantee Ticket please see the Event Page.
> The "Refine Guarantee Ticket" can even be used to refine gears that normally cannot be refined.
- Updated dialog for the Paradise Group Quests.
- Added a cute pet related NPC.
- Changed it so you can no longer use the Neutralizer while using Warg/Magic Gear/Dragon Riding.
- Corrected the effect for the Arch Bishop Skill "Expiatio".
- Corrected an issue where when equipping specific headgear, the damage from Throw Shuriken was abnormally large.
- Corrected an issue with the probability of success for the Raido Rune.
Sakray Test Server
- Monster Status/Experience/Element/Element level's have been adjusted on some monsters.
- Corrected an issue where you couldn't advance the "Attitude to the new world" quest after starting the "Atonement to the new world" quest.
- Changed the Banshee Master Card wording (suffix -> prefix). (Of Exorcism -> Surplus).
- The following Rune Knight skills have been adjusted.
Enchant Blade: Attack Power increased. The user's base level and INT influence the attack power gained.
Sonic Wave: The user's base level influences the attack power. The skill's damage no longer decreases with the distance from the caster. Cooldown time changed to 2s.
Hundred Spear: The user's base level influences the attack power. Variable cast time changed to 1s. Skill aftercast delay changed to 0.5s.
Wind Cutter: The user's base level influences the attack power. Variable casting time removed. Cooldown time changed to 2s. Skill aftercast delay removed.
Ignition Break: The user's base level influences the attack power. The medium ranged damage is increased. Fixed cast time is removed, variable cast time set to 1s. Cooldown time changed to 2s. Skill aftercast delay removed.
Dragon Training: When mounted on a dragon and using a spear class weapon, you ignore all size modifiers doing 100% damage to all targets. While mounted, the skill level of spear mastery increases your attack power. Your attack speed without Dragon Training is now 75%.
Dragon Breath: MaxSP now influences the damage more than before. DFixed cast time set to 0.5s. The variable cast time based on skill level.
Phantom Thrust: Attack power increased. The spear mastery skill now influences the damage. The user's base level influences the attack power.
Rune Mastery: When manufacturing rune stones, you can create 1~2 at a time with level 5, and 1~3 with level 10. When doing this, if it would put you over your maximum number of carried runes, the rune stone is consumed but the ingredients are not used.
Nathuz Rune (Refresh): Duration increased from 10s to 30s. Can be used while under the wanderer/minstrel status "Deep Sleep", and removes the Deep Sleep status.
Wyrd Rune (Storm Blast): The user's base level influences the attack power. Fixed cast time is removed, variable cast time set to 2s. Cooldown time is removed. Taking damage no longer interrupts the skill.
Othila Rune (Fighting Spirit): An additional effect is added to the rune's current effect. When using the rune stone, the user gains some attack speed. This effect is influenced by your level of Rune Mastery, with higher levels of Rune Mastery adding more attack speed. | s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579251799918.97/warc/CC-MAIN-20200129133601-20200129163601-00239.warc.gz | CC-MAIN-2020-05 | 4,092 | 35 |
http://blog.seanbonner.com/2008/12/02/links-for-2008-12-01/ | code | In The Public Domain: Enclosing the Commons of the Mind James Boyle introduces readers to the idea of the public domain and describes how it is being tragically eroded by our current copyright, patent, and trademark laws.
Most of the time I live in Los Angeles, though I spend a lot of time in Tokyo too. I've run hackerspaces and blog networks, an art gallery, design firm and a record label. I'm one of the co-founders of Safecast, and currently act as Global Director. I'm a Shuttleworth Fellow, an MIT researcher and sit on the board of CicLAvia. | s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187826642.70/warc/CC-MAIN-20171023202120-20171023222120-00017.warc.gz | CC-MAIN-2017-43 | 550 | 2 |
https://developer.2c2p.com/docs/download-merchant-server-setup | code | Merchant Server Setup
For the 2C2P PGW SDK to work, a server-side implementation is needed. Refer to the guide below for instructions to set up a merchant server.
- PHP (v2.0.0): Download
How to integrate:
Refer to: Get payment token and Inquiry payment result guideline.
Updated 3 months ago | s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224657720.82/warc/CC-MAIN-20230610131939-20230610161939-00782.warc.gz | CC-MAIN-2023-23 | 292 | 6 |
http://evidencesynthesishackathon.com/software/eviatlas.html | code | Systematic Maps are, according to the Environmental Evidence Journal, “overviews of the quantity and quality of evidence in relation to a broad (open) question of policy or management relevance.” In simple terms, this means that documents are categorized according to the type, location, and publication information available for each work within a particular topic. Systematic maps are often used for environmental research, where it is particularly important to track the location of study sites. The spatial nature of a systematic map, particularly for environmental research, means that academics often use some kind of geographic map to analyze and present their information. Understanding the academic community’s familiarity with the R programming language, the team decided to build a webapp using R Shiny that could automate certain parts of creating a systematic map for environmental research.
Using EviAtlas, a researcher will be able to shorten the time needed to generate key plots and maps of the previous work they’re analyzing. What might have taken a full days work in the past could now be condensed into about an hour. EviAtlas is still in development, and is open to pull requests. | s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912203529.38/warc/CC-MAIN-20190324230359-20190325012359-00499.warc.gz | CC-MAIN-2019-13 | 1,210 | 2 |
https://forum.netgate.com/topic/145174/openvpn-over-gateway-group-doesn-t-fail-back/2?lang=en-US | code | OpenVPN over Gateway group doesn't fail back
I am having the following issue with an OpenVPN client. I am running the lastest version of pfsense: 2.4.4-RELEASE-p3
I have a multi-WAN setup (WAN + 4G) and a gateway group FAILOVER_WAN configured as follows:
WAN - Tier 1 - Monitor IP: the public IP used as gateway for the connection
WAN_PPP - Tier 2 - Monitor IP: 18.104.22.168
Trigger - Packet loss or high latency (not relevant to the problem)
Outbound NAT disabled since I only need the routes received from the server.
For testing purposes, all interfaces have Permit on any protocol, any source, any destination in Firewall.
OpenVPN client using the FAILOVER_WAN as a source interface, connected to an OpenVPN server (not relevant to the issue, either).
If I disconnect the WAN, the gateway goes down and the OpenVPN service restarts and uses the 4G connection (Tier 2). All good so far.
The issue comes when I reconnect the WAN, the gateway status changes to UP, but the OpenVPN connection stays up over the 4G connection. A simple service restart solves the issue and OpenVPN client connects via the Tier 1 GW.
Any suggestion on how to make the OpenVPN client fail back to Tier 1 would be appreciated.
I have tried (seen on other discussions on this forum):
Use sticky connections - enabled/disabled
Flush all states when a gateway goes down - enabled/disabled
Default gateway IPv4 - Automatic/None/WAN_GW/WAN_PPP_GW
Added openvpn in Service Watchdog
The only workaround I can think of is adding a cron script to restart the service every hour or so, but it's not ellegant at all.
This post is deleted!
Thanks for the fast reply.
It is 100% set as such, since otherwise it wouldn't fail Tier1->Tier2
The issue is that OpenVPN doesn't switch back to Tier1 once it is available again (the system detects the GW back up, but I guess openvpn process needs to be restarted for it to reconnect using the preffered interface.)
I just found this client settings item in your post. (And deleted my answer)
It is so strange, I use completely the same scenario (ex. trigger) and my OVPN client reconnect back after main GW became available.
UPDATE: might have found the issue.
Tier 1 WAN interface was set up to receive IP via DHCP. I have changed this to static IP/SM/GW and now failover works great in both ways :) . Don't know what the underlying issue is. To be fair, the DHCP server I was using was kind of slow. It would lease the IP after >30 sec, I will do some more digging.
@robertko huh! I'm not sure that dhcp mode is the reason itself (my pf wan interfaces based on dhcp both) But slow dhcp lease process could affect restoring process. It is great that you solved it! Congrats!) | s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030334942.88/warc/CC-MAIN-20220926211042-20220927001042-00533.warc.gz | CC-MAIN-2022-40 | 2,686 | 27 |
https://premium.wpmudev.org/forums/topic/marketpress-multisite-user-shop-product-photos | code | Marketpress Multisite User Shop Product Photos.
Right now Users are only able to upload one main photo. Other photos are generated within the about product section and added under there, which I do not like.
I would like to have additional photo widget where it automatically sizes the photo and only allows users to add 5 photos per item.
What is the best solution? Is the main photo widget/section able to be modified? Or can I duplicate that section and direct it as additional images in code? | s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027317516.88/warc/CC-MAIN-20190822215308-20190823001308-00536.warc.gz | CC-MAIN-2019-35 | 496 | 4 |
http://de.mathworks.com/help/symbolic/mupad_ref/groebner-stronglyindependentsets.html?nocookie=true | code | Strongly independent set of variables
This functionality does not run in MATLAB.
groebner::stronglyIndependentSets(G) computes a strongly independent set of variables modulo the ideal generated by G.
A set of variables S is strongly independent modulo an ideal I if no leading term of an element of the Gröbner basis of I consists entirely of elements of S. A set is maximally strongly independent if no proper superset of it is strongly independent. Two maximally strongly independent set may be of different size.
groebner::stronglyIndependentSets accepts Gröbner bases in the format returned by groebner::gbasis.
The following example has been given by Moeller and Mora in 1983.
G:=map([X0^8*X2, X0*X3, X1^8*X3, X1^7*X3^2, X1^6*X3^3, X1^5*X3^4, X1^4*X3^5, X1^3*X3^6, X1^2*X3^7, X1*X3^8], poly, [X3, X2, X1, X0]): groebner::stronglyIndependentSets(G)
The Gröbner basis of an ideal: a list.
List of the form [d, S, M], where d is an integer equal to the dimension of the ideal generated by G, S is the greatest strongly independent set of variables, and M is a set consisting of all maximal strongly independent sets of variables or a piecewise consisting of such lists.
Kredel H. and V. Weispfenning, "Computing dimension and independent sets for polynomial ideals", JSC volume 6 (1988), 231-247. | s3://commoncrawl/crawl-data/CC-MAIN-2015-11/segments/1424936463104.7/warc/CC-MAIN-20150226074103-00181-ip-10-28-5-156.ec2.internal.warc.gz | CC-MAIN-2015-11 | 1,302 | 10 |
https://medicine.iu.edu/research-centers/alzheimers/About/Organization-and-Core-Structure/neuroimaging | code | The IADRC Neuroimaging Core of the Indiana Alzheimer’s Disease Research Center serves as a regional and national resource for the aging and neurodegeneration research community—spanning academia, health care and pharmaceutical industry. The core provides access to and collaborative support for the application of advanced neuroimaging in clinical and translational research.
The IADRC Neuroimaging Core is developing and implementing advanced imaging protocols and imaging-genetics analysis techniques, collaborating with NIA-funded initiatives, including Alzheimer’s Disease Genetics Consortium (ADGC), Alzheimer’s Disease Neuroimaging Initiative (ADNI), Dominantly Inherited Alzheimer’s Network (DIAN), the Alzheimer’s Disease Cooperative Study (ADCS).
- Support funded research in the IADRC, IU Center for Aging Research, and related programs that could benefit from advanced neuroimaging
- Provide standardized, state-of-the-art neuroimaging and analysis protocols
- Expand transdisciplinary regional neuroscience research using advanced neuroimaging tools to study disease mechanisms and treatments for neurodegeneration
- Support and collaborate with major national and international AD-related research consortia using neuroimaging and genetics methods
- Provide educational opportunities for scientists at all levels from undergraduates to post-doctoral fellows and faculty, as well as dissemination to the community | s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882572089.53/warc/CC-MAIN-20220814234405-20220815024405-00454.warc.gz | CC-MAIN-2022-33 | 1,437 | 7 |
https://www.stonefieldsoftware.com/OnlineHelp/alere/_0oq0rmfsd.htm | code | This is an advanced topic intended for programmers.
You can use a "script" file to programmatically customize Stonefield Query. After Stonefield Query sets up its data environment (which it does only one time after a particular database has been chosen), it looks for a file named Setup.sqs and executes the code in that file if it exists.
To create a script file, create a text file called Setup.sqs in the directory where Stonefield Query's data files are located. Edit this file using any text editor (not a word processor, such as Microsoft Word, which stores binary files, but an editor such as Notepad that stores text files).
Although script code can be used for anything, the most common use is to change information in the Stonefield Query data dictionary. Here's an example. This one script Stonefield Query to treat a couple of fields as numeric and date values rather than the character data they're actually stored as.
loField = SQApplication.DataEngine.Fields.Item('MACHINERY.CAPACITY') loField.Type = 'N' loField.OutputLength = 5 loField.Expression = 'val(MACHINERY.CAPACITY)' loField.Picture = '99,999' loField = SQApplication.DataEngine.Fields.Item('PRODUCTS.LASTDATE') loField.Type = 'D' loField.OutputLength = 8 loField.Expression = 'ctod(PRODUCTS.LASTDATE)'
The first line in this code asks the fields collection to return a field object for the CAPACITY field (which contains the desired value) of the MACHINERY table. The code then changes the field data type to numeric ("N"), the output length to 5, the output expression to the expression that converts a character value to a numeric value (the VAL() function), and sets the picture to "99,999" so numbers are formatted with commas. The code then does the same thing for the LASTDATE field in the PRODUCTS table, using the appropriate changes for a date field ("D" instead of "N," a width of 8, and CTOD(), which converts a character to a date, rather than VAL() for the output expression).
In previous versions of Stonefield Query, scripting was used to create your own calculated fields. However, that's no longer necessary because you can use the Formula Editor to do that.
© Stonefield Software Inc., 2022 • Updated: 02/01/21
Comment or report problem with topic | s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224648322.84/warc/CC-MAIN-20230602040003-20230602070003-00022.warc.gz | CC-MAIN-2023-23 | 2,244 | 9 |
https://studiopulpit.com/2020/02/13/making-ai-work-with-small-data/ | code | Making AI work with small data
In a recent MAPI survey 58% of the respondents indicated that the lack of (good quality) data is the main barrier to deployment of AI. In many industries data scientists and machine learning engineers need to deal with small data sets rather than big data. In this article in Industryweek you can find an excellent overview of the techniques available how to deal with small data sets. To me this is a must read for anyone who works in data science/ analytics. | s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817222.1/warc/CC-MAIN-20240418160034-20240418190034-00724.warc.gz | CC-MAIN-2024-18 | 491 | 2 |
http://www.cplusplus.com/user/mekki/ | code | print on paperi coded a simple program to store data and would like to add a "print out" option. my idea is to loa...
print on paperhi,
like to print on physical paper. i use xcode.
thanks a lot for any idea,
graphicthx a lot,
i am about to simulate the solar system - the math is done but i have no idea how to convert my...
how to search string and ignore cases?that will keep me busy,
thx a lot,
This user does not accept Private Messages | s3://commoncrawl/crawl-data/CC-MAIN-2015-14/segments/1427131300773.77/warc/CC-MAIN-20150323172140-00225-ip-10-168-14-71.ec2.internal.warc.gz | CC-MAIN-2015-14 | 441 | 9 |
http://scripting.com/stories/2010/09/01/whatsProducedAtHackathons.html | code | I've never participated in a hackathon, so I can't tell whether they produce anything or not. I'm assuming that at least some of the readers of this blog have been to one or more.
What kind of software is produced?
Were any commercial products hatched at hackathons? Were any of them successful?
My intuition says that they're pointless exercises. I have at times gone on retreats with the idea of emerging with something useful or marketable, but I have to say the big leaps in software that I've seen or taken part of, come with steady daily work, when you've built up a head of steam over weeks or months.
It seems to me that the hackathon idea is more a dream of investors or marketers, that they can get a bunch of programmers in a room to invent something they can make money with. But that's just my impression. | s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224646257.46/warc/CC-MAIN-20230531022541-20230531052541-00571.warc.gz | CC-MAIN-2023-23 | 818 | 5 |
https://forum.duolingo.com/comment/19297781/Nu-nu-vorbi%C8%9Bi-engleza | code | yes... yes it should... but if you put a "!" after "engleza" it's correct because it's an order, because it puts emphasis on what not to speak... not on just speaking
i guess it's like 'the english' apparently you could say: limba engleză or engleza
"No, do not speak English" and "No, you do not speak English" are very different sentences yet they are both accepted as answers. Why?
I think that it's because it's supposed to be sort of saying that you aren't to speak English. Like, No! Don't speak English, you! Please correct me if I'm wrong. | s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195526237.47/warc/CC-MAIN-20190719115720-20190719141720-00393.warc.gz | CC-MAIN-2019-30 | 548 | 4 |
https://sourceforge.net/directory/os%3Aindependent/os%3Awinnt/environment%3Awin32/?page=7 | code | This is a very nice and simple client with GUI to connect to Net4India Broadband Server. This works on Linux/Windows. It also minimizes to notification area. Hover mouse over status icon in notification area and you will get session information.
*Updated 6-11-2010 * Nova is a remote assistance tool developed in C# .NET 3.5. Using Nova, you can administrate remote computers through a File Explorer, Registry Editor, Process Manager, Live View, and Chat.
NovaGrid is a light spreadsheet made with Python / Tkinter, based on Tktable. The objective of this project is to have a light spreasheet python object which can be used into larger projects. The code is totaly written in Python (actualy 2.2.2).
Pattern recognition software package. It includes several classification and clustering algorithms. It can read data from a set of images, an ASCII file or a JDBC connection. A small TCP data server with its corresponding JDBC driver is included.
OOIMapping is an OpenOffice.org Add-On assisting the user in generating Documents that comply to Information MappingŽ standards. Information Mapping is a registered trademark of Information Mapping, Inc
wxPython based macro installer/manager for OpenOffice.org. Run on Linux and Windows 2000.
ORM.NET : Object-Relational Mapping and Code Generation Tool for .NET
The Object Library Documentation Project defines standards and utilities for creating documentation structures for object models and packages like ActiveX Dynamic Libraries, Controls, .NET Assemblies, Java packages, and web services.
Objirc Chat is a Java based IRC client. This 5th generation product includes: * a cool Swing based GUI (designed for real people not just Java developers) * event-based support for the IRC protocol * Python and Java based plugin support * DCC pr
ObscureEncrypt is a high level security encrypter. It encrypts at three levels, 448 bit Blowfish encryption, Obscurity, high level encryption. Making it almost impossible to decrypt
Office in a Box is a easy to use, easy to maintain and highly configurable document server for workgroups and small to medium sized companies.
Open Direct Connect is a multi-protocol file sharing program that supports the protocol used by Direct Connect ( http://www.neo-modus.com/ ).
Open LivePerson is the new name for the project CLCSWI (Customer Live Chat System - Web Interface). The purpose of this project is to make an opensource system based on most features from HumanClick, LivePerson, PHPLiveSupport, etc.
OMS is used for sequencing events on a timeline, allowing control of software and hardware for music, video, and demoscene productions. It is fairly similar to trackers, which are music editors, famous for their use in games and demoscene productions.
OpenGo is a program and framework for working on AI through software that plays the game of Go. As programming Go is difficult, OpenGo helps by provding commonly used code and user interface so that the engine and algorithms can be focused on.
Projeto de ferramenta OpenSource para Gerência/Análise de Riscos. Implementação de módulos para conformidade com normas, como ISO, COBIT, integração com ferramenta de inventário (GLPI), Gerência de PSI.
A new generation of trakkers based on portable open-source engine and a set of user-friendly front-ends. You don't like existing trakkers? Create your one with existing engine!
A 3-D Othello (Reversi) gaming environment.
PUMz.News help you clipping web sites for mobile devices. It converts form web sites such as bulletin boards, forums, news to minimized web sites for mobile devices. It runs on Windows XP, Vista or above. It requires .NET 2.0 Framework.
Pears is a three-pane newsfeed (RSS/RDF/Atom) aggregator which caches downloaded feeds for offline use. It has a clean, uncluttered interface, it's easy to use and works on Windows, Linux and MacOSX. You can extend its functionality with plugins.
PgWorksheet is a simple GUI frontend to PostgreSQL for executing SQL queries and psql commands without using the psql command line tool. Please see the project home page for bug reports, forums and installation instructions.
A linguistic tool to aid in the study of Linguistics/Phonology, specifically distinctive features of possible language sounds. Comprised of both a Visual C++ .NET version as well as a Java based web applet version. The C++ version has all but been ab
A new portal written in php with content managment (MySQL).
PopED (Population Experimental Design) is a software tool for computing optimal experimental designs. The software has been developed with an emphasis on drug trials based on population models (non-linear mixed effects models). | s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084889660.55/warc/CC-MAIN-20180120142458-20180120162458-00630.warc.gz | CC-MAIN-2018-05 | 4,669 | 24 |
http://webapps.stackexchange.com/questions/tagged/instapaper+rss | code | Web Applications Meta
to customize your list.
more stack exchange communities
Start here for a quick overview of the site
Detailed answers to any questions you might have
Discuss the workings and policies of this site
Instapaper RSS feed is not working?
I have been using Instapaper for a long time. I have used RSS feed from it and Google Reader for feed reader. Somehow this stopped working for about a month ago. I don't get any feeds from Instapaper ...
Oct 19 '11 at 18:38
newest instapaper rss questions feed
Hot Network Questions
Super/Krazy Glue on Outdoor Gear
Nonpaid, volunteering position in a lab
What artist drew this image, that was used on the Warriors of the Eternal Sun Genesis Game (not about the game)
Can someone please explain buffers to me?
Avoiding the Hermann grid illusion
Are patches a bad sign for the customer?
Is it OK to use while ((line = r.readLine()) != null) construct?
Flammable or Inflammable
Doing all the simulations and plotting, do I deserve an authorship?
Why are most people not persuaded by rational arguments?
Initialize List<> with some count of elements
If DOS is single-tasking, how was multitasking possible in old version of Windows?
Connecting Circular Rooms
Can this recursive binary search in C be made more concise?
Flatten a nGon face to avoid face normal's deformation
4, 8, 15, 16, 23, 42
Calculate a tip
Are the orcs of the D&D core canon not above eating sentient humanoids?
Undefined control sequence (\email, \address) in combine class
Where does Allah come from?
2D collision detection
Should I learn to use LaTex to write up a History Masters Thesis?
sum and product of two rational numbers are both integers
How can I kill myself?
more hot questions
Life / Arts
Culture / Recreation
TeX - LaTeX
Unix & Linux
Ask Different (Apple)
Geographic Information Systems
Science Fiction & Fantasy
Seasoned Advice (cooking)
Personal Finance & Money
English Language & Usage
Mi Yodeya (Judaism)
Cross Validated (stats)
Theoretical Computer Science
Meta Stack Overflow
Stack Overflow Careers
site design / logo © 2014 stack exchange inc; user contributions licensed under
cc by-sa 3.0 | s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1394678664178/warc/CC-MAIN-20140313024424-00079-ip-10-183-142-35.ec2.internal.warc.gz | CC-MAIN-2014-10 | 2,136 | 53 |
https://hktsoft.net/trees-in-java/ | code | Every computer user who has worked with a hierarchical file system has seen tree displays. Of course, directories and files are only one of the many examples of tree-like organizations. Many tree structures arise in everyday life, such as the hierarchy of countries, states, and cities shown in Figure 11.11.
As programmers, we often need to display tree structures. Fortunately, the Swing library has a JTree class for this purpose. The JTree class (together with its helper classes) takes care of laying out the tree and processing user requests for expanding and collapsing nodes. In this section, you will learn how to put the JTree class to use.
As with the other complex Swing components, we must focus on the common and useful cases and cannot cover every nuance. If you want to achieve something unusual, we recommend that you consult Graphic Java™, Third Edition, by David M. Geary or Core Swing by Kim Topley.
Before going any further, let’s settle on some terminology (see Figure 11.12). A tree is composed of nodes. Every node is either a leaf or it has child nodes. Every node, with the exception of the root node, has exactly one parent. A tree has exactly one root node. Sometimes you have a collection of trees, each with its own root node. Such a collection is called a forest.
1. Simple Trees
In our first example program, we will simply display a tree with a few nodes (see Figure 11.14). As with many other Swing components, you need to provide a model of the data, and the component displays it for you. To construct a JTree, supply the tree model in the constructor:
TreeModet model = …;
var tree = new JTree(model);
How do you obtain a tree model? You can construct your own model by creating a class that implements the TreeModel interface. You will see later in this chapter how to do that. For now, we will stick with the DefaultTreeModel that the Swing library supplies.
To construct a default tree model, you must supply a root node.
TreeNode root = . . .;
var model = new DefaultTreeModel(root);
TreeNode is another interface. Populate the default tree model with objects of any class that implements the interface. For now, we will use the concrete node class that Swing supplies—namely, DefaultMutableTreeNode. This class implements the MutableTreeNode interface, a subinterface of TreeNode (see Figure 11.13).
A default mutable tree node holds an object—the user object. The tree renders the user objects for all nodes. Unless you specify a renderer, the tree displays the string that is the result of the toString method.
In our first example, we use strings as user objects. In practice, you would usually populate a tree with more expressive user objects. For example, when displaying a directory tree, it makes sense to use File objects for the nodes.
You can specify the user object in the constructor, or you can set it later with the setUserObject method.
var node = new DefaultMutableTreeNode(“Texas”);
Next, you need to establish the parent/child relationships between the nodes. Start with the root node and use the add method to add the children:
var root = new DefaultMutableTreeNode(“World”);
var country = new DefaultMutableTreeNode(“USA”);
var state = new DefaultMutableTreeNode(“California”);
Figure 11.14 illustrates how the tree will look.
Link up all nodes in this fashion. Then, construct a DefaultTreeModel with the root node. Finally, construct a JTree with the tree model.
var treeModel = new DefaultTreeModel(root);
var tree = new JTree(treeModel);
Or, as a shortcut, you can simply pass the root node to the JTree constructor. Then the tree automatically constructs a default tree model:
var tree = new JTree(root);
Listing 11.8 contains the complete code.
When you run the program, the tree first looks as in Figure 11.15. Only the root node and its children are visible. Click on the circle icons (the handles) to open up the subtrees. The line sticking out from the handle icon points to the right when the subtree is collapsed and down when the subtree is expanded (see Figure 11.16). We don’t know what the designers of the Metal look-and-feel had in mind, but we think of the icon as a door handle. You push down on the handle to open the subtree.
You can use the following magic incantation to turn off the lines joining parents and children (see Figure 11.18):
Conversely, to make sure that the lines are shown, use
Another line style, “Horizontal”, is shown in Figure 11.19. The tree is displayed with horizontal lines separating only the children of the root. We aren’t quite sure what it is good for.
By default, there is no handle for collapsing the root of the tree. If you like, you can add one with the call
Figure 11.20 shows the result. Now you can collapse the entire tree into the root node.
Conversely, you can hide the root altogether. You will thus display a forest—a set of trees, each with its own root. You still must join all trees in the forest to a common root; then, hide the root with the instruction
Look at Figure 11.21. There appear to be two roots, labeled “USA” and “Germany.” The actual root that joins the two is made invisible.
Let’s turn from the root to the leaves of the tree. Note that the leaves have an icon different from the other nodes (see Figure 11.22).
When the tree is displayed, each node is drawn with an icon. There are actually three kinds of icons: a leaf icon, an opened nonleaf icon, and a closed nonleaf icon. For simplicity, we refer to the last two as folder icons.
The node renderer needs to know which icon to use for each node. By default, the decision process works like this: If the isLeaf method of a node returns true, then the leaf icon is used; otherwise, a folder icon is used.
The isLeaf method of the DefauttMutabteTreeNode class returns true if the node has no children. Thus, nodes with children get folder icons, and nodes without children get leaf icons.
Sometimes, that behavior is not appropriate. Suppose we added a node “Montana” to our sample tree, but we’re at a loss as to what cities to add. We would not want the state node to get a leaf icon because, conceptually, only the cities are leaves.
The JTree class has no idea which nodes should be leaves. It asks the tree model. If a childless node isn’t automatically a conceptual leaf, you can ask the tree model to use a different criterion for leafiness—namely, to query the “allows children” node property.
For those nodes that should not have children, call
Then, tell the tree model to ask the value of the “allows children” property to determine whether a node should be displayed with a leaf icon. Use the setAsksAUowsChildren method of the DefaultTreeModel class to set this behavior:
With this decision criterion, nodes that allow children get folder icons, and nodes that don’t allow children get leaf icons.
Alternatively, if you construct the tree from the root node, supply the setting for the “asks allows children” property in the constructor.
var tree = new JTree(root, true); // nodes that don’t allow children get teat icons
1.1 Editing Trees and Tree Paths
In the next example program, you will see how to edit a tree. Figure 11.23 shows the user interface. If you click the Add Sibling or Add Child button, the program adds a new node (with title New) to the tree. If you click the Delete button, the program deletes the currently selected node.
To implement this behavior, you need to find out which tree node is currently selected. The JTree class has a surprising way of identifying nodes in a tree. It does not deal with tree nodes but with paths of objects, called tree paths. A tree path starts at the root and consists of a sequence of child nodes (see Figure 11.24).
You might wonder why the JTree class needs the whole path. Couldn’t it just get a TreeNode and keep calling the getParent method? In fact, the JTree class knows nothing about the TreeNode interface. That interface is never used by the TreeModet interface; it is only used by the DefauttTreeModet implementation. You can have other tree models in which the nodes do not implement the TreeNode interface at all. If you use a tree model that manages other types of objects, those objects might not have getParent and getChitd methods. They would of course need to have some other connection to each other. It is the job of the tree model to link nodes together. The JTree class itself has no clue about the nature of their linkage. For that reason, the JTree class always needs to work with complete paths.
The TreePath class manages a sequence of Object (not TreeNode!) references. A number of JTree methods return TreePath objects. When you have a tree path, you usually just need to know the terminal node, which you can get with the getLastPathComponent method. For example, to find out the currently selected node in a tree, use the getSetectionPath method of the JTree class. You will get a TreePath object back, from which you can retrieve the actual node.
TreePath setectionPath = tree.getSetectionPath();
var setectedNode = (DefauttMutabteTreeNode) setectionPath.getLastPathComponent();
Actually, since this particular query is so common, there is a convenience method that gives the selected node immediately:
var setectedNode = (DefauttMutabteTreeNode) tree.getLastSetectedPathComponent();
This method is not called getSetectedNode because the tree does not know that it contains nodes—its tree model deals only with paths of objects.
Once you have the selected node, you can edit it. However, do not simply add children to a tree node:
setectedNode.add(newNode); // No!
If you change the structure of the nodes, you change the model but the associated view is not notified. You could send out a notification yourself, but if you use the insertNodeInto method of the DefauttTreeModet class, the model class takes care of that. For example, the following call appends a new node as the last child of the selected node and notifies the tree view:
modet.insertNodeInto(newNode, setectedNode, setectedNode.getChitdCount());
The analogous call removeNodeFromParent removes a node and notifies the view:
If you keep the node structure in place but change the user object, you should call the following method:
The automatic notification is a major advantage of using the DefauttTreeModet. If you supply your own tree model, you have to implement automatic notification by hand. (See Core Swing by Kim Topley for details.)
When the view is notified of a change in the node structure, it updates the display but does not automatically expand a node to show newly added children. In particular, if a user in our sample program adds a new child node to a node for which children are currently collapsed, the new node is silently added to the collapsed subtree. This gives the user no feedback that the command was actually carried out. In such a case, you should make a special effort to expand all parent nodes so that the newly added node becomes visible. Use the makeVisible method of the JTree class for this purpose. The makeVisible method expects a tree path leading to the node that should become visible.
Thus, you need to construct a tree path from the root to the newly inserted node. To get a tree path, first call the getPathToRoot method of the DefauttTreeModet class. It returns a TreeNode array of all nodes from a node to the root node. Pass that array to a TreePath constructor.
For example, here is how you make the new node visible:
TreeNode nodes = model.getPathToRoot(newNode);
var path = new TreePath(nodes);
Now, suppose your tree is contained inside a scroll pane. After the tree node expansion, the new node might still not be visible because it falls outside the viewport. To overcome that problem, call
instead of calling makeVisibte. This call expands all nodes along the path and tells the ambient scroll pane to scroll the node at the end of the path into view (see Figure 11.25).
By default, tree nodes cannot be edited. However, if you call
the user can edit a node simply by double-clicking, editing the string, and pressing the Enter key. Double-clicking invokes the default cell editor, which is implemented by the DefauttCettEditor class (see Figure 11.26). It is possible to install other cell editors, using the same process that you have seen in our discussion of table cell editors.
Listing 11.9 shows the complete source code of the tree editing program. Run the program, add a few nodes, and edit them by double-clicking. Observe how collapsed nodes expand to show added children and how the scroll pane keeps added nodes in the viewport.
2. Node Enumeration
Sometimes you need to find a node in a tree by starting at the root and visiting all children until you have found a match. The DefauttMutabteTreeNode class has several convenience methods for iterating through nodes.
The breadthFirstEnumeration and depthFirstEnumeration methods return enumeration objects whose nextElement method visits all children of the current node, using either a breadth-first or depth-first traversal. Figure 11.27 shows the traversals for a sample tree—the node labels indicate the order in which the nodes are traversed.
Breadth-first enumeration is the easiest to visualize. The tree is traversed in layers. The root is visited first, followed by all of its children, then the grandchildren, and so on.
To visualize depth-first enumeration, imagine a rat trapped in a tree-shaped maze. It rushes along the first path until it comes to a leaf. Then, it backtracks and turns around to the next path, and so on.
Computer scientists also call this postorder traversal because the search process visits the children before visiting the parents. The postOrderEnumeration method is a synonym for depthFirstEnumeration. For completeness, there is also a preOrderEnumeration, a depth-first search that enumerates parents before the children.
Here is the typical usage pattern:
Enumeration breadthFirst = node.breadthFirstEnumeration();
do something with breadthFirst.nextElement();
Finally, a related method, pathFromAncestorEnumeration, finds a path from an ancestor to a given node and enumerates the nodes along that path. That’s no big deal— it just keeps calling getParent until the ancestor is found and then presents the path in reverse order.
In our next example program, we put node enumeration to work. The program displays inheritance trees of classes. Type the name of a class into the text field on the bottom of the frame. The class and all of its superclasses are added to the tree (see Figure 11.28).
In this example, we take advantage of the fact that the user objects of the tree nodes can be objects of any type. Since our nodes describe classes, we store Class objects in the nodes.
We don’t want to add the same class object twice, so we need to check whether a class already exists in the tree. The following method finds the node with a given user object if it exists in the tree:
public DefaultMutableTreeNode findUserObject(Object obj)
Enumeration e = root.breadthFirstEnumeration(); while (e.hasMoreElements())
DefaultMutableTreeNode node = (DefaultMutableTreeNode) e.nextElement();
3. Rendering Nodes
In your applications, you will often need to change the way a tree component draws the nodes. The most common change is, of course, to choose different icons for nodes and leaves. Other changes might involve changing the font of the node labels or drawing images at the nodes. All these changes are possible by installing a new tree cell tenderer into the tree. By default, the JTree class uses DefaultTreeCettRenderer objects to draw each node. The DefaultTreeCettRenderer class extends the JLabet class. The label contains the node icon and the node label.
You can customize the display in three ways.
- You can change the icons, font, and background color used by a DefaultTreeCettRenderer. These settings are used for all nodes in the tree.
- You can install a renderer that extends the DefaultTreeCettRenderer class and vary the icons, fonts, and background color for each node.
- You can install a renderer that implements the TreeCellRenderer interface to draw a custom image for each node.
Let us look at these possibilities one by one. The easiest customization is to construct a DefaultTreeCellRenderer object, change the icons, and install it into the tree:
var renderer = new DefaultTreeCellRenderer();
renderer.setLeafIcon(new ImageIcon(“blue-ball.gif”)); // used for leaf nodes
renderer.setClosedIcon(new ImageIcon(“red-ball.gif”)); // used for collapsed nodes
renderer.setOpenIcon(new ImageIcon(“yellow-ball.gif”)); // used for expanded nodes
You can see the effect in Figure 11.28. We just use the “ball” icons as placeholders—presumably your user interface designer would supply you with appropriate icons to use for your applications.
We don’t recommend that you change the font or background color for an entire tree—that is really the job of the look-and-feel.
However, it can be useful to change the font of some nodes in a tree to highlight them. If you look carefully at Figure 11.28, you will notice that the abstract classes are set in italics.
To change the appearance of individual nodes, install a tree cell renderer. Tree cell renderers are very similar to the list cell renderers we discussed earlier in this chapter. The TreeCettRenderer interface has a single method:
Component getTreeCettRendererComponent(JTree tree, Object value, boolean selected, boolean expanded,
boolean leaf, int row, boolean hasFocus)
The getTreeCeURendererComponent method of the DefaultTreeCellRenderer class returns this—in other words, a label. (The DefaultTreeCellRenderer class extends the JLabel class.) To customize the component, extend the DefauttTreeCettRenderer class. Override the getTreeCettRendererComponent method as follows: Call the superclass method so it can prepare the label data, customize the label properties, and finally return this.
class MyTreeCettRenderer extends DefauttTreeCettRenderer
public Component getTreeCeURendererComponent(JTree tree, Object value, boolean selected, boolean expanded, boolean leaf, int row, boolean hasFocus)
Component comp = super.getTreeCeURendererComponent(tree, value, selected,
expanded, leaf, row, hasFocus);
DefauttMutabteTreeNode node = (DefauttMutabteTreeNode) vatue;
look at node.getUserObject();
Font font = appropriate font;
The ClassNameTreeCeURenderer in Listing 11.10 sets the class name in either the normal or italic font, depending on the ABSTRACT modifier of the Ctass object. We don’t want to set a particular font because we don’t want to change whatever font the look-and-feel normally uses for labels. For that reason, we use the font from the label and derive an italic font from it. Recall that only a single shared JLabet object is returned by all calls. We need to hang on to the original font and restore it in the next call to the getTreeCettRendererComponent method.
Also, note how we change the node icons in the CtassTreeFrame constructor.
4. Listening to Tree Events
Most commonly, a tree component is paired with some other component. When the user selects tree nodes, some information shows up in another window. See Figure 11.29 for an example. When the user selects a class, the instance and static variables of that class are displayed in the text area to the right.
To obtain this behavior, you need to install a tree selection listener. The listener must implement the TreeSetectionListener interface—an interface with a single method:
void vatueChanged(TreeSetectionEvent event)
That method is called whenever the user selects or deselects tree nodes. Add the listener to the tree in the normal way:
You can specify whether the user is allowed to select a single node, a contiguous range of nodes, or an arbitrary, potentially discontiguous, set of nodes. The JTree class uses a TreeSelectionModel to manage node selection. You need to retrieve the model to set the selection state to one of SINGLE_TREE_SELECTION, CONTIGUOUS_TREE_SELECTION, or DISCONTIGUOUS_TREE_SELECTION. (Discontiguous selection mode is the default.) For example, in our class browser, we want to allow selection of only a single class:
int mode = TreeSelectionModel.SINGLE_TREE_SELECTION;
Apart from setting the selection mode, you need not worry about the tree selection model.
To find out the current selection, query the tree with the getSelectionPaths method:
TreePath selectedPaths = tree.getSelectionPaths();
If you restricted the user to single-item selection, you can use the convenience method getSelectionPath which returns the first selected path or null if no path was selected.
Listing 11.10 shows the frame class for the class tree program. The program displays inheritance hierarchies and customizes the display to show abstract classes in italics. (See Listing 11.11 for the cell renderer.) You can type the name of any class into the text field at the bottom of the frame. Press the Enter key or click the Add button to add the class and its superclasses to the tree. You must enter the full package name, such as java.util.ArrayList.
This program is a bit tricky because it uses reflection to construct the class tree. This work is done inside the addClass method. (The details are not that important. We use the class tree in this example because inheritance yields a nice supply of trees without laborious coding. When you display trees in your applications, you will have your own source of hierarchical data.) The method uses the breadth-first search algorithm to find whether the current class is already in the tree by calling the findUserObject method that we implemented in the preceding section. If the class is not already in the tree, we add the superclasses to the tree, then make the new class node a child and make that node visible.
When you select a tree node, the text area to the right is filled with the fields of the selected class. In the frame constructor, we restrict the user to singleitem selection and add a tree selection listener. When the vatueChanged method is called, we ignore its event parameter and simply ask the tree for the current selection path. As always, we need to get the last node of the path and look up its user object. We then call the getFietdDescription method which uses reflection to assemble a string with all fields of the selected class.
5. Custom Tree Models
In the final example, we implement a program that inspects the contents of an object, just like a debugger does (see Figure 11.30).
Before going further, compile and run the example program. Each node corresponds to an instance field. If the field is an object, expand it to see its instance fields. The program inspects the contents of the frame window. If you poke around a few of the instance fields, you should be able to find some familiar classes. You’ll also gain some respect for how complex the Swing user interface components are under the hood.
What’s remarkable about the program is that the tree does not use the DefauttTreeModet. If you already have data that are hierarchically organized, you might not want to build a duplicate tree and worry about keeping both trees synchronized. That is the situation in our case—the inspected objects are already linked to each other through the object references, so there is no need to replicate the linking structure.
The TreeModet interface has only a handful of methods. The first group of methods enables the JTree to find the tree nodes by first getting the root, then the children. The JTree class calls these methods only when the user actually expands a node.
int getChitdCount(Object parent)
Object getChild(Object parent, int index)
This example shows why the TreeModet interface, like the JTree class itself, does not need an explicit notion of nodes. The root and its children can be any objects. The TreeModet is responsible for telling the JTree how they are connected.
The next method of the TreeModet interface is the reverse of getChitd:
int getIndexOfChitd(Object parent, Object chitd)
Actually, this method can be implemented in terms of the first three—see the code in Listing 11.12.
The tree model tells the JTree which nodes should be displayed as leaves:
boolean isLeaf(Object node)
If your code changes the tree model, the tree needs to be notified so it can redraw itself. The tree adds itself as a TreeModetListener to the model. Thus, the model must support the usual listener management methods:
void addTreeModelListener(TreeModelListener l)
void removeTreeModetListener(TreeModetListener t)
You can see the implementations for these methods in Listing 11.13.
When the model modifies the tree contents, it calls one of the four methods of the TreeModetListener interface:
void treeNodesChanged(TreeModetEvent e)
void treeNodesInserted(TreeModelEvent e)
void treeNodesRemoved(TreeModelEvent e)
void treeStructureChanged(TreeModelEvent e)
The TreeModelEvent object describes the location of the change. The details of assembling a tree model event that describes an insertion or removal event are quite technical. You only need to worry about firing these events if your tree can actually have nodes added and removed. In Listing 11.12, we show how to fire one event by replacing the root with a new object.
Finally, if the user edits a tree node, your model is called with the change:
void valueForPathChanged(TreePath path, Object newValue)
If you don’t allow editing, this method is never called.
If you don’t need to support editing, constructing a tree model is easily done. Implement the three methods:
int getChildCount(Object parent)
Object getChild(Object parent, int index)
These methods describe the structure of the tree. Supply routine implementations of the other five methods, as in Listing 11.12. You are then ready to display your tree.
Now let’s turn to the implementation of the example program. Our tree will contain objects of type Variable.
For example, suppose you inspect the variable
That variable has a type Employee.class, a name “joe”, and a value—the value of the object reference joe. In Listing 11.14, we define a class Variable that describes a variable in a program:
var v = new Variable(Employee.class, “joe”, joe);
If the type of the variable is a primitive type, you must use an object wrapper for the value.
new Variable(double.class, “salary”, new Double(salary));
If the type of the variable is a class, the variable has fields. Using reflection, we enumerate all fields and collect them in an ArrayList. Since the getFields method of the Class class does not return the fields of the superclass, we need to call getFields on all superclasses as well. You can find the code in the Variable constructor. The getFields method of our Variable class returns the array of fields. Finally, the toString method of the Variable class formats the node label. The label always contains the variable type and name. If the variable is not a class, the label also contains the value.
Let’s move on to the tree model. The first two methods are simple.
public Object getRoot()
public int getChildCount(Object parent)
return ((Variable) parent).getFields().size();
The getChild method returns a new Variable object that describes the field with the given index. The getType and getName methods of the Field class yield the field type and name. By using reflection, you can read the field value as f.get(parentValue). That method can throw an IllegalAccessException. However, we made all fields accessible in the Variable constructor, so this won’t happen in practice.
Here is the complete code of the getChild method:
public Object getChild(Object parent, int index)
ArrayList fields = ((Variable) parent).getFields();
var f = (Field) fields.get(index);
Object parentValue = ((Variable) parent).getValue(); try
return new Variable(f.getType(), f.getName(), f.get(parentValue));
catch (IllegalAccessException e)
These three methods reveal the structure of the object tree to the JTree component. The remaining methods are routine—see the source code in Listing 11.13.
There is one remarkable fact about this tree model: It actually describes an infinite tree. You can verify this by following one of the WeakReference objects. Click on the variable named referent. It leads you right back to the original object. You get an identical subtree, and you can open its WeakReference object again, ad infinitum. Of course, you cannot store an infinite set of nodes; the tree model simply generates the nodes on demand as the user expands the parents. Listing 11.12 shows the frame class of the sample program.
Source: Horstmann Cay S. (2019), Core Java. Volume II – Advanced Features, Pearson; 11th edition. | s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224652959.43/warc/CC-MAIN-20230606150510-20230606180510-00498.warc.gz | CC-MAIN-2023-23 | 28,797 | 183 |
https://onmsft.com/news/new-blockchain-development-kit-released-for-microsofts-azure/ | code | Earlier this year, Microsoft dedicated a not insignificant amount of stage time at their developer conference to explain blockchain technology. While much of the section went over my head at the time, in the subsequent months, I’ve seen several companies introduce new apps and migration services to leverage evolving blockchain technologies.
On that note, Microsoft just released a new software development kit on its Azure cloud platform that will allow businesses to create their own distributed ledger technology (DLT) to work with SaaS solutions and third parties over the course of days, rather than the months it used to take.
Microsoft’s new DLT solution is called Azure Blockchain Workbench and was released yesterday November 20, 2018, to commercial users, but ABW has been tested and implemented for some time by other Asia-Pacific companies such as China Binary Sale Technology and High Cloud.
More specifically, yesterday’s release is focusing on features and tools to interface with mobile clients, Internet of Things (IoT), SMS and ambient computing systems among other technologies. In addition to the release of ABW, Microsoft also launched a set of extension tools in Logic App and Flow Connectors to boost workflows around blockchain networks.
As Marc Mercuri, a Microsoft principal programmer at Blockchain Engineering puts it, “The resulting applications will run atop a network that has higher rated cloud performance than other large-scale providers and enable federating identities between participants using Azure Active Directory.” | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100602.36/warc/CC-MAIN-20231206162528-20231206192528-00274.warc.gz | CC-MAIN-2023-50 | 1,567 | 5 |
https://blog.jetbrains.com/dotnet/2007/04/18/resharper-30-eap-in-action-the-todo-explorer/ | code | Resharper 3.0 EAP in Action – The Todo Explorer
The ReSharper 3.0 EAP has been in progress for about two weeks now. This is the first post in a series that will look in detail at some of the new features that will be available in this latest version of ReSharper.
The To-do Explorer
In Visual Studio.NET 2005, you can use the Task List window to show comments with the keyword todo in it. However, this only works with files that are open in the IDE; it’s not going to show you TODOs in other parts of your solution that you don’t currently have open, or any other keyword EXCEPT todo.
The Visual Studio.NET 2005 Task List window
In ReSharper 3.0, you get the much more useful To-do Explorer. When you open this window (found in ReSharper -> Windows -> To-do Explorer), by default you see a solution-wide listing of all comments with the prefix NOTE, TODO or BUG.
The new To-do Explorer window
The default keywords that the To-do explorer looks for can be changed or added to by clicking the “To-do settings” icon in the window.
In the settings window, you can create new keywords to search for or edit the existing keywords as you’d like. This could be useful, for example, if you have a references to JIRA tickets in your comments, like:
//PROJ-1655: The following code block was added because of a change request from a client
You could add a pattern to match the keyword “
PROJ-####” and see all JIRA tickets mentioned in your code.
Editing keywords in the To-do explorer
These To-dos can also be exported to a text file and can also be grouped to make the To-dos easier to read, so you can find which projects, namespaces, directories, etc … contain To-dos.
The To-do Explorer view options
Subscribe to Blog updates
Thanks, we've got you!
Eager, Lazy and Explicit Loading with Entity Framework Core
Entity Framework Core (EF Core) supports a number of ways to load related data. There’s eager loading, lazy loading, and explicit loading. Each of these approaches have their own advantages and drawbacks. In this post, let’s have a quick look at each of these ways to load data for navigational prope…
OSS Power-Ups: bUnit – Webinar Recording
The recording of our webinar, OSS Power-Ups: bUnit, with Egil Hansen and Steven Giesel, is available. This was the twelfth episode of our OSS Power-Ups series, where we put a spotlight on open-source .NET projects. Subscribe to our community newsletter to receive notifications about future webi…
Accelerating Your Testing Workflow with Unit Test Creation and Navigation
Unit tests play an important role in our daily development workflow. They help us ensure our codebase's correctness when writing new functionality or performing refactorings to improve readability and maintainability. In the process, we often create new test files that accompany the p…
Introducing Predictive Debugging: A Game-Changing Look into the Future
With the introduction of debugging tools, software developers were empowered to interactively investigate the control flow of software programs to find bugs in live environments. At JetBrains, we've always strived to improve the art of debugging. Besides the more standard things you expect from a de… | s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233510300.41/warc/CC-MAIN-20230927135227-20230927165227-00662.warc.gz | CC-MAIN-2023-40 | 3,209 | 25 |
https://www.experts-exchange.com/questions/24576471/192-168-22-X-on-public-internet.html | code | Trying to figure out why a client's internet is slow-did a tracert from there to a few webservers, and noticed that after packets are handed off from office router to Covad, it passes thru routers with 192.168.22.122 and 192.168.22.73 addresses. (As a routine matter, when setting up client networks, I use different numbers for the third octet in each case so that I can connect via VPN if at one client and get an emergency call from another). Seems to me that the above numbers should NOT be used by Covad equipment! Just fortunate that I haven't used .22. on any other clients, or I might have run into problems earlier. Can someone confirm that the Covad numbering above should NOT be used on the public internet? | s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267155676.21/warc/CC-MAIN-20180918185612-20180918205612-00448.warc.gz | CC-MAIN-2018-39 | 718 | 1 |
https://newcircus.dk/welcome-to-my-new-blog/ | code | It has been very difficult for me to keep my homepage updated with news and photos.
Now it should be a little bit easier with this new blog.
Currently I am working at the amusement park Dyrehavsbakken 7 days a week.
I do solo shows in the main stage every day, and juggling and mime around the park.
Except from that I have some time for fitness, office, parrot and accordion. | s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296816893.9/warc/CC-MAIN-20240414192536-20240414222536-00895.warc.gz | CC-MAIN-2024-18 | 376 | 5 |
http://financialcryptography.com/mt/archives/000530.html | code | Voting is a particularly controversial application (or feature) for FC because of the difficulty in both setting the requirements, and the 'political requirement' of ensuring a non-interfered vote for each person. I've just got back from an alpine retreat where I participated in a small experiment to test votes with tokens, called Beetle In A Box. The retreat was specifically purposed to do the early work to build a voting application, so it was fun to actually try out some voting.
Following on from our pressed flower production technique, we 'pressed & laminated' about 100 'beatles,' or symbols looking like squashed beatles. These were paired in plus and minus form, and created in sets of similar symbols, with 5 colours for different purposes. Each person got a set of 10, being 5 subsets of two complementary pairs each.
The essence of the complicated plus and minus tokens was to try out delegated voting. Each user could delegate their plus token to another user, presumably on the basis that this other user would know more and was respected on this issue. But they could always cast their minus to override the plus, if they changed their minds. More, it is a sad fact of voting life that unless you are in Australia, where political voting is compulsory, most people don't bother to turn up anyway.
To simulate this, we set up 4 questions (allocating 4 colours) to be held at 4 different places - a deliberate conflict. One of the questions was the serious issue of naming the overall project and we'd been instructed to test that; the others were not so essential. Then we pinned up 21 envelopes for all the voters and encouraged people to put their plus tokens in the named envelope of their delegatee.
When voting time came, chaos ensued. Many things went wrong, but out of all that we did indeed get a vote on the critical issue (not that this was considered binding in any way). Here's the stats:
Number of direct voters: 4 Number of delegated votes: 3 Therefore, total votes cast: 7 Winning project name: Mana, with 3 votes.
So, delegated voting increased the participation by 75%, taking total participation to 33% (7 of 21 participants). That's significant - that's a huge improvement over the alternate and indicates that delegated voting may really be useful or even needed. But, another statistic indicates there is a lot more that we could have done:
Number of delegated votes, not cast: 9
That is, in the chaos of the game, many more people delegated their votes, but the tokens didn't make it to the ballot. The reasons for this are many: one is just the chaos of understanding the rules and the way they changed on the fly! Another is that many delegatees simply didn't participate at all, and in particular the opinion leaders who collected fat envelopes forgot they were supposed to vote, and just watched the madness around them (in increasing frustration!).
Canny FCers will recognise another flaw in the design - having placed the tokens into envelopes, the delegators then had to become delegatees and collect from their envelopes. And, if they were not to then attend that meeting (there were 4 conflicting meetings, recall) then the delegatees would become delegators again and re-delegate. Thus forcing the cycle to start again, ad infinitum.
Most people only went to the pinboard once. So the formal delegation system simply failed on efficiency grounds, and it is lucky that some smart political types did some direct swaps and trades on their delegated votes.
How then to do this with physical tokens is an open question. If one wants infinite delegation, I can't see how to do it efficiently. With a computer system, things become more plausible, but even then how do we model a delegated vote in software?
Is it a token? Not quite, as the delegate vote can be overridden and thus we need a token that can be yanked back. Or overridden or redirected. So it could be considered to be an accounting entry - like nymous account money - but even then, the records of a payment from alice to bob need to be reversable.
One final result. Because I was omnipresent (running the meeting that took the important vote) I was able to divine which were the delegated votes. And, in this case, if the delegated votes had been stripped out, and only direct voting handled, the result of the election decision would have changed: the winning name would have been Medici, which was what I voted for.
Which I count as fairly compelling evidence that whatever the difficulties in implementing delegated voting, it's a worthwhile goal for the future. | s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917122720.81/warc/CC-MAIN-20170423031202-00260-ip-10-145-167-34.ec2.internal.warc.gz | CC-MAIN-2017-17 | 4,576 | 15 |
https://swimm.io/learn/legacy-code/legacy-code-refactoring-before-during-and-after-refactoring | code | What Does It Mean to Refactor Legacy Code?
Refactoring legacy code involves restructuring the existing codebase without changing its external behavior. It’s about improving the code’s structure and design, making it cleaner, more efficient, and easier to maintain.
Legacy code is often associated with older software, but it can be any code that is difficult to understand, maintain, or extend. It’s the type of code that developers often dread working with because it’s complex, confusing, and usually lacks adequate documentation or tests.
Refactoring legacy code is a critical practice in software development. It’s not just about cleaning up the mess. It’s about turning a liability into an asset, transforming a burden into a reliable foundation for future development.
Why Refactor Legacy Code?
There are several compelling reasons to refactor legacy code. Let’s explore a few of the most important ones.
Improving Maintainability and Readability
One of the main reasons for refactoring legacy code is to improve its maintainability. Legacy code is notoriously difficult to maintain. It’s often written in a convoluted manner, without clear structure or meaningful variable names. By refactoring the code, we can improve its readability, making it easier for developers to understand and work with the code.
Another key reason to refactor legacy code is to enhance its performance. Over time, codebases can become bloated and inefficient, leading to slow execution times and poor user experience. Through refactoring, we can streamline the code, eliminating unnecessary complexity and redundancies, and ultimately improving the software’s performance.
Addressing Security Vulnerabilities
Legacy code often contains security vulnerabilities that can be exploited by malicious actors. These vulnerabilities may be due to outdated programming practices, dependencies on insecure libraries, or simply errors in the code. Refactoring provides an opportunity to review the code and address these vulnerabilities, improving the software’s security.
Facilitating New Features and Integrations
Refactoring legacy code also facilitates future feature additions or integrations. A clean, well-structured codebase is easier to extend and integrate with other systems. It provides a solid foundation for future development, enabling the team to add new features or integrate the software with other systems more efficiently and with fewer risks.
Reducing Long-Term Costs
Finally, refactoring legacy code can reduce long-term costs. While there is an upfront cost associated with refactoring, the long-term savings can be significant. A clean, efficient codebase requires less maintenance, reduces the risk of bugs, and improves team productivity. Over time, these benefits translate into lower costs and higher return on investment.
Refactoring Legacy Code: Pre-Refactoring Steps
Before you dive into refactoring legacy code, there are a few critical steps you need to take to set yourself up for success.
Understanding the Codebase
The first step in refactoring legacy code is to understand the codebase. This includes understanding the code’s structure, functionality, and dependencies. It’s also crucial to understand the business logic and requirements that the code is intended to fulfill. This understanding will guide your refactoring efforts and help you avoid inadvertently breaking the existing functionality.
Learn more in our detailed guide to legacy codebase (coming soon)
Setting Up a Version Control System
A version control system is a must-have tool for any software development project, and it’s especially important when refactoring legacy code. It allows you to track changes to the codebase, making it easier to identify and rollback changes that cause issues. It also facilitates collaboration, enabling multiple developers to work on the codebase without stepping on each other’s toes.
Creating a Testing Safety Net
Before you start refactoring, you should ensure you have a robust suite of automated tests in place. These tests will help you verify that the refactored code still behaves as expected, preventing regressions. With a robust test suite in place, you can confidently make changes, knowing that if your changes have unexpected side effects, the tests will catch them.
Refactoring Legacy Code: Strategies for Effective Refactoring
Incremental changes mean that instead of rewriting large chunks of code all at once, you should focus on making small, manageable changes one at a time.
This minimizes the risk of introducing new bugs into the system. When you make small changes, it is easier to test that part of the code and ensure that it is working as expected. Furthermore, if you do introduce a bug, it is easier to identify and fix.
Another advantage of making incremental changes is that it allows you to gradually understand the codebase. Legacy code can often be complex and poorly documented. By making small changes, you can slowly build up your understanding of the code, which will make the refactoring process easier in the long run.
Dependencies are parts of the code that rely on other parts, on third-party libraries, or external systems to function correctly. If these dependencies are not properly managed, they can create a tangled web of code that is difficult to understand and maintain.
When refactoring, it is important to identify and isolate these dependencies. This can be done by creating interfaces or abstract classes that encapsulate the dependency, making it easier to manage and modify. This also makes the code more modular and easier to test, as you can test each dependency in isolation.
Furthermore, isolating dependencies can make the code more robust and less prone to bugs. If a change is made to one part of the code, it is less likely to affect other parts of the code if the dependencies are well managed.
Prioritizing High-Impact Areas
When refactoring, it can be tempting to try and fix everything at once. However, this is often not feasible or practical. Instead, you should focus on the areas of the code that will have the most impact.
These high-impact areas are typically the parts of the code that are used most frequently, or that are causing the most problems. By focusing your efforts on these areas, you can make the most significant improvements to the codebase.
To identify these high-impact areas, you can use tools like code coverage reports, performance profiling, and bug reports. These tools can help you identify the parts of the code that are used most frequently, or that are causing the most issues.
Using Refactoring Patterns
There are many different refactoring patterns available, each with its own advantages and disadvantages. Some of the most common patterns include Extract Method, Move Method, Rename Method, and Replace Conditional with Polymorphism.
Using refactoring patterns can help to standardize the refactoring process, making it easier to understand and follow. They can also help to improve the quality of the code, by encouraging good coding practices and reducing complexity.
Refactoring Legacy Code: Post-Refactoring Considerations
Validating Existing Functionality
After you’ve refactored your legacy code, the first post-refactoring consideration is to validate that the functionality remains intact. This means testing the code to ensure that it still works as expected.
This is where unit tests and integration tests come in handy. These tests can help you to verify that the functionality of the code has not been compromised during the refactoring process. If any issues are found, they can be quickly identified and fixed.
Additionally, code reviews can also be a valuable tool for ensuring that the functionality remains intact. By having another developer review your changes, they can provide a fresh perspective and may spot potential issues that you may have overlooked.
Monitoring Performance and Other Metrics
There are many different metrics that you can monitor, including response times, memory usage, and CPU usage. This has two aspects: ensuring the application’s performance has not deteriorated as a result of refactoring, and measuring the success of refactoring in improving performance. By comparing the performance of the application before and after the refactoring, you can determine whether the changes have had a positive or negative impact.
Gathering Feedback from End-Users or Stakeholders
Gathering feedback from end-users or stakeholders can provide valuable insights into how the refactoring has impacted the user experience.
Feedback can be gathered through surveys, interviews, or user testing sessions. This feedback can help you to identify any issues that may have arisen as a result of the refactoring, or any areas that could be improved further.
Legacy Code Documentation with Swimm
Refactoring legacy code is a practice that goes beyond merely tidying up messy code; it’s about turning the burden of legacy code into a solid foundation for future development. Refactoring paves the way for improved maintainability, enhanced performance, security fortification, and the seamless integration of new features. The long-term benefits in reduced costs and increased productivity make it a strategic investment for any organization.
Swimm is a dev tool designed to make legacy code documentation easier and more accessible. In essence, it makes the entire refactoring process easier. Swimm’s code-coupled documentation, combined with its easy discoverability in the IDE, makes it an ideal companion for your refactoring efforts. With Swimm, you can ensure that your documentation remains up to date as your code changes, improving your team’s productivity and code quality. | s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817382.50/warc/CC-MAIN-20240419074959-20240419104959-00585.warc.gz | CC-MAIN-2024-18 | 9,764 | 51 |
https://resource.dopus.com/t/restrict-flat-view-subfolders/15088/13 | code | I'm in England, although it's difficult to say what my effective timezone is.
It won't work in everything, but it will work in some other things.
[quote]If I have 3 folders which I do not want to include, I guess the following would be the filter string if the folders were named Trip1_Jan, Trip2_Feb, and Trip3_Mar, and they were the only folders whose name started with "Trip".
You may need brackets, and you should not include the .* as the folder names do not have dots or extensions in them.
So this is what you could use:
Using that as a filter in Flat View may have unwanted effects, though. It will also filter out any files with names like Trip01.jpg
You'd be better off using
That is safer, since you're unlikely to have files called any of those three names.
[quote]And, as you noted, to include these 3 extensions, it would be
To exclude then
You probably need the extra brackets again:
That gets quite difficult if you're trying to do that in a single wildcard expression.
To do complex filtering, filtering some things in and other things out, I would use the Find tool (Tools -> Find Panel) instead of trying to use filtering and Flat View. With the Find tool you can break things down and create complex rules much more easily. | s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882573760.75/warc/CC-MAIN-20220819191655-20220819221655-00174.warc.gz | CC-MAIN-2022-33 | 1,243 | 13 |
http://fortifyitinc.blogspot.com/2015/ | code | Everyone please have online backups and backups to external drives because I have had two client servers get hit with CryptoLocker 3.0, the variant of CryptoWall. Luckily my backup worked and are back to normal.
Please make sure you have multiple backups! There are a lot of online services out there.
Friday, March 13, 2015
Fortify I.T. received Best of Lawrenceville 2015 for a SECOND consecutive year! I'm really excited about winning this again as it is a testament for how hard I work for my clients. Thank you everyone for the support! | s3://commoncrawl/crawl-data/CC-MAIN-2018-34/segments/1534221216475.75/warc/CC-MAIN-20180820140847-20180820160847-00640.warc.gz | CC-MAIN-2018-34 | 541 | 4 |
http://hyderabad-blues1.blogspot.com/2012/02/hm251486-help-regarding-blackberry.html | code | can v track a lost blackberry bold phone if the thief puts in his/her sim.
Can the police get the information of the new sim and thru that the owner of sim.
A friend of mine recently lost blackberry bold phone. he lodged a police FIR as well, but nothing seems to happen..and nothing will.
I want to know that whether it is possible if the thief inserts his sim into the cellphone and the service provider could get notified about the stolen cell anyhow..
Also tell that could this phone be tracked or traced, i read the forums and everywhere it was like if already an application to track it was installed. Is there any other way??
Plz help me in this regard
Amity Business School | s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267860168.62/warc/CC-MAIN-20180618090026-20180618110026-00215.warc.gz | CC-MAIN-2018-26 | 681 | 7 |
http://sar-roo.ru/web-2-0-for-dating-13350.html | code | The Web 2.0 Suicide Machine is a service that helps users tired of My Space, Linked In and Twitter, to "commit suicide in social networks", by automatically "removing their private content and friend relationships" (but without deleting or deactivating their accounts).Rather than deleting user accounts, it removes private content and friendships.
First, Web 2.0 doesn't refer to a specific advance in Web technology.
Instead, Web 2.0 refers to a set of techniques for Web page design and execution.
Second, some of these techniques have been around since the World Wide Web first launched, so it's impossible to separate Web 1.0 and Web 2.0 in a time line.
The definition of Web 1.0 completely depends upon the definition of Web 2.0.
However, Web 2.0 does not refer to a specific version of the Web, but rather a series of technological improvements. | s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400209665.4/warc/CC-MAIN-20200923015227-20200923045227-00100.warc.gz | CC-MAIN-2020-40 | 852 | 6 |
https://www.califmom.com/home/2011/11/13/things-i-have-learned.html | code | Unless asked, don't tell your significant other what to wear. This person is, presumably, an adult. Get over your issues (yes, they're YOUR issues) with image.
Life is too short to worry about how the dishwasher is loaded. If somebody has been kind enough to load the dishwasher, say thank you.
Giving other people ultimatums rarely ends well. Same thing goes for pointing out a laundry-list of their faults. Ultimately, the only person you can fix/control/change is you. Start there and magical things will happen. The other person might still be the same, but you'll probably be a lot happier. Check the mirror, fix yourself, and move on. So. Much. Easier.
That's all the wisdom I've got for today, kids. Need to go conquer some Operation Eleanor fears and help my man build a bed. | s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514575484.57/warc/CC-MAIN-20190922094320-20190922120320-00323.warc.gz | CC-MAIN-2019-39 | 783 | 4 |
https://www.aging-us.com/figure/202340/f7 | code | Figure 7. GO enrichment analysis (A) and KEGG pathway analysis (B) of potential targets of aging and constipation. The size of the bubbles in each bubble chart represents the gene counts of this entry. The colors from cold to warm represent the P values from large to small. Each bubble chart is sorted by P value. GO: gene ontology; KEGG: kyoto encyclopedia of genes and genomes. (C) The target-pathway network diagram of aging and constipation. The yellow labels and the blue labels represent disease targets and signing pathways, respectively. The red labels and the green labels represent the common signaling pathways between aging and constipation, and the unique disease targets for aging or constipation, respectively. The size of each label represents its degree. The thickness and color of the lines represent edge. The edges represent the interactions between them and node sizes are proportional to their degree. The blue nodes and the red nodes represent targets pathways, respectively. | s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030337855.83/warc/CC-MAIN-20221006191305-20221006221305-00448.warc.gz | CC-MAIN-2022-40 | 999 | 1 |
http://animaticons.co/about/ | code | A better set of animated icons
As a web designer, developer, and animator for the past 12 years, I’m always looking for ways to make my work more engaging. One of the best ways is through motion and animation. But adding animation to a website is not always easy:
- SVG animations can take dozens, even hundreds, of lines of code to implement.
- Video often comes with a large file size, and doesn’t automatically play on mobile devices, requiring the user to interact before they see any motion.
- Flash and Silverlight animations require third-party plugins, and they’re not supported by some smartphones.
That’s why I developed Animaticons as a set of high-resolution animated GIFs that you can customize.
Animated GIFs are:
- Small in file size.
- Compatible with all major browsers, emails, and smartphones.
- Play automatically without user interaction.
- Don’t require any special plugins or libraries.
I hope you enjoy the icons and find creative ways to use them in your projects. There are more icons coming in the future. Check back for updates or sign up to get notified when new icons are released. | s3://commoncrawl/crawl-data/CC-MAIN-2017-47/segments/1510934806419.21/warc/CC-MAIN-20171121151133-20171121171133-00476.warc.gz | CC-MAIN-2017-47 | 1,121 | 12 |
http://kalvimalar.dinamalar.com/news-details.asp?id=226&cat=10&q=General | code | At the outset, you must be interested in cricket, must have the experience of playing cricket during school and college days.
Above all, you must be familiar with the basic rules of cricket. Apart from this, only with an experience in umpiring during the matches held in your locality, you can obtain umpiring skills. All cricket associations in the state conduct examinations to assess the umpiring skill and knowledge in cricket. One can acquire umpiring skill only after umpiring in the district matches and state matches, and then move on step by step to test matches, one day matches and 20- 20 matches. | s3://commoncrawl/crawl-data/CC-MAIN-2013-48/segments/1386163041955/warc/CC-MAIN-20131204131721-00015-ip-10-33-133-15.ec2.internal.warc.gz | CC-MAIN-2013-48 | 608 | 2 |
https://gfxdomain.co/2018/05/udemy-advanced-game-programming-in-unity-3d-120931.html | code | In this course you will learn advanced game programming in unity 3d.
What Will I Learn?
Use debugger better
work on advanced scripting
Create dynamic terrain
Create custom GUI controls
Work on networking
Optimize your game
Basic Understanding of computer is required.
Unity is a cross-platform game engine developed by Unity Technologies,which is primarily used to develop both three-dimensional and two-dimensional video games and simulations for computers, consoles, and mobile devices. First announced only for OS X at Apple’s Worldwide Developers Conference in 2005, it has since been extended to target 27 platforms.
Unity is the number one game engine, the favorite of game developers working on iOS apps and console games alike. This course picks up the programming side of Unity. Instructor concentrates on scripting, custom GUI controls, prefab customization, multiplayer networking features like chat, and optimizations that will make your game play better on mobile devices and desktops alike. Take a deep dive into Unity’s controls and learn how to build more interactive, high-performance 3D games.
if you will have any question you can add a discussion and i will answer it as soon as possible. At the end of the course, if you will not be satisfied you can get a complete refund. I hope you will enroll, see you inside.
Who is the target audience?
People who want to learn video game development in unity 3D.
Udemy – Advanced Game Programming in Unity 3D.part1.rar
Udemy – Advanced Game Programming in Unity 3D.part2.rar
Udemy – Advanced Game Programming in Unity 3D.part3.rar
Udemy – Advanced Game Programming in Unity 3D.part4.rar
Udemy – Advanced Game Programming in Unity 3D.part5.rar
Udemy – Advanced Game Programming in Unity 3D.part6.rar
Udemy – Advanced Game Programming in Unity 3D.part7.rar | s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764500080.82/warc/CC-MAIN-20230204012622-20230204042622-00564.warc.gz | CC-MAIN-2023-06 | 1,833 | 21 |
http://freecode.com/tags/ocr?page=1&sort=created_at&with=&without=8824 | code | GlyphViewer is a desktop application that allows users to build translations from text in images and export them into different image formats or even HTML. Users can use OCR technology to identify text in images, such as English, German, Chinese, Arabic, Japanese, and many more. A unique feature of the application is its support for Ancient Egyptian Hieroglyphs.
Character Recognition is an Android app that allows the user to take a photo (or use existing image files on the device) and then apply the Tesseract OCR engine to extract the text in the photo. It is currently supporting English text, but other language support will be added in the future.
getxbook is a collection of tools to download books from websites. There are tools to download from Google Books' "book preview", Amazon's "look inside the book", and Barnes and Noble's "book viewer". There is an optional GUI written in Tcl/Tk, and some shell scripts using OCR to create plain text or searchable PDFs and DjVu files from the downloaded books.
Aspose.OCR for .NET is a character recognition component built to allow developers to add OCR functionality in their ASP .NET Web applications, Web services, and applications. It provides a simple set of classes for controlling character recognition tasks and supports BMP and TIFF.
MALODOS helps you to scan, store, and easily retrieve all your personal documents. Its storage format is open and documented, so your document archive can remain accessible even without MALODOS. The documents themselves are stored as standard PDF files, while their metadata (such as title, tags, and description) are stored into a separate SQLite database in an open format. With MALODOS, you can also manage existing files in PDF, JPEG, TIFF, and other formats, so you can still use the documents that you've already scanned. You can connect to any external OCR program to give access to a fulltext search feature.
Mayan EDMS is a document manager Web application with custom metadata indexing, file serving integration, and OCR capabilities. It features user defined metadata fields, dynamic default values for metadata, lookup support for metadata, filesystem integration by means of metadata indexing directories, user defined document UUID generation, local file or server side staging file uploads, batch uploading of many documents with the same metadata, user defined document checksum algorithms, previews for a great deal of image formats including PDF, document OCR and searching, automatic grouping of documents by metadata, permissions and roles support, multi-page document support, page transformations, distributed OCR processing, and support for multiple languages. | s3://commoncrawl/crawl-data/CC-MAIN-2014-15/segments/1397609533121.28/warc/CC-MAIN-20140416005213-00073-ip-10-147-4-33.ec2.internal.warc.gz | CC-MAIN-2014-15 | 2,683 | 6 |
https://rosadohill.com/products/mary-python-genuine-leather-messenger-bag | code | LOW STOCK WARNING
Carry your belongings in style with this Mary Python Genuine Leather Messenger Bag. Featuring a snake pattern and a luxurious design, this bag will become the new treasure of your wardrobe. You will feel fabulous and gorgeous and get everyone’s attention with this bag. Now available in white print, snake print, gold green, and leopard print. | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100327.70/warc/CC-MAIN-20231202042052-20231202072052-00083.warc.gz | CC-MAIN-2023-50 | 363 | 2 |
https://www.databricks.com/blog/2016/07/28/continuous-applications-evolving-streaming-in-apache-spark-2-0.html | code | Since its release, Spark Streaming has become one of the most widely used distributed streaming engines, thanks to its high-level API and exactly-once semantics. Nonetheless, as these types of engines became common, we’ve noticed that developers often need more than just a streaming programming model to build real-time applications. At Databricks, we’ve worked with thousands of users to understand how to simplify real-time applications. In this post, we present the resulting idea, continuous applications, which we have started to implement through the Structured Streaming API in Apache Spark 2.0.
Most streaming engines focus on performing computations on a stream: for example, one can map a stream to run a function on each record, reduce it to aggregate events by time, etc. However, as we worked with users, we found that virtually no use case of streaming engines only involved performing computations on a stream. Instead, stream processing happens as part of a larger application, which we’ll call a continuous application. Here are some examples:
- Updating data that will be served in real-time. For instance, developers might want to update a summary table that users will query through a web application. In this case, much of the complexity is in the interaction between the streaming engine and the serving system: for example, can you run queries on the table while the streaming engine is updating it? The “complete” application is a real-time serving system, not a map or reduce on a stream.
- Extract, transform and load (ETL). One common use case is continuously moving and transforming data from one storage system to another (e.g. JSON logs to an Apache Hive table). This requires careful interaction with both storage systems to ensure no data is duplicated or lost -- much of the logic is in this coordination work.
- Creating a real-time version of an existing batch job. This is hard because many streaming systems don’t guarantee their result will match a batch job. For example, we've seen companies that built live dashboards using a streaming engine and daily reporting using batch jobs, only to have customers complain that their daily report (or worse, their bill!) did not match the live metrics.
- Online machine learning. These continuous applications often combine large static datasets, processed using batch jobs, with real-time data and live prediction serving.
These examples show that streaming computations are part of larger applications that include serving, storage, or batch jobs. Unfortunately, in current systems, streaming computations run on their own, in an engine focused just on streaming. This leaves developers responsible for the complex tasks of interacting with external systems (e.g. managing transactions) and making their result consistent with the the rest of the application (e.g., batch jobs). This is what we'd like to solve with continuous applications.
We define a continuous application as an end-to-end application that reacts to data in real-time. In particular, we’d like developers to use a single programming interface to support the facets of continuous applications that are currently handled in separate systems, such as query serving or interaction with batch jobs. For example, here is how we would handle the use cases above:
- Updating data that will be served in real time. The developer would write a single Spark application that handles both updates and serving (e.g. through Spark’s JDBC server), or would use an API that automatically performs transactional updates on a serving system like MySQL, Redis or Apache Cassandra.
- Extract, transform and load (ETL). The developer would simply list the transformations required as in a batch job, and the streaming system would handle coordination with both storage systems to ensure exactly-once processing.
- Creating a real-time version of an existing batch job. The streaming system would guarantee results are always consistent with a batch job on the same data.
- Online machine learning. The machine learning library would be designed to combine real-time training, periodic batch training, and prediction serving behind the same API.
The figure below shows which concerns are usually handled in streaming engines, and which would be needed in continuous applications:
Structured Streaming is a new high-level API we have contributed to Apache Spark 2.0 to support continuous applications. It is, first, a higher-level API than Spark Streaming, bringing in ideas from the other structured APIs in Spark (DataFrames and Datasets)—most notably, a way to perform database-like query optimizations. More importantly, however, Structured Streaming also incorporates the idea of continuous applications to provide a number of features that no other streaming engines offer.
- Strong guarantees about consistency with batch jobs. Users specify a streaming computation by writing a batch computation (using Spark’s DataFrame/Dataset API), and the engine automatically incrementalizes this computation (runs it continuously). At any point, the output of the Structured Streaming job is the same as running the batch job on a prefix of the input data. Most current streaming systems (e.g. Apache Storm, Kafka Streams, Google Dataflow and Apache Flink) do not provide this "prefix integrity" property.
- Transactional integration with storage systems. We have taken care in the internal design to process data exactly once and update output sinks transactionally, so that serving applications always see a consistent snapshot of the data. While the Spark 2.0 release only supports a few data sources (HDFS and S3), we plan to add more in future versions. Transactional updates were one of the top pain points for users of Spark and other streaming systems, requiring manual work, so we are excited to make these part of the core API.
- Tight integration with the rest of Spark. Structured Streaming supports serving interactive queries on streaming state with Spark SQL and JDBC, and integrates with MLlib. These integrations are only beginning in Spark 2.0, but will grow in future releases. Because Structured Streaming builds on DataFrames, many other libraries of Spark will naturally run over it (e.g., all feature transformations in MLlib are written against DataFrames).
Apart from these unique characteristics, Structured Streaming has other new features to simplify streaming, such as explicit support for “event time” to aggregate out of order data, and richer support for windowing and sessions. Achieving its consistency semantics in a fault-tolerant manner is also not easy—see our sister blog post about the API and execution model.
Structured Streaming is still in alpha in Spark 2.0, but we hope you try it out and send feedback. Our team and many other community members will be expanding it in the next few releases.
As a simple example of Structured Streaming, the code below shows an Extract, Transform and Load (ETL) job that converts data from JSON into Apache Parquet. Note how Structured Streaming simply uses the DataFrame API, so the code is nearly identical to a batch version.
[col xs="12" md="6"]
// Read JSON continuously from S3 logsDF = spark.readStream.json("s3://logs") // Transform with DataFrame API and save logsDF.select("user", "url", "date") .writeStream.parquet("s3://out") .start()
[col xs="12" md="6"]
// Read JSON once from S3 logsDF = spark.read.json("s3://logs") // Transform with DataFrame API and save logsDF.select("user", "url", "date") .write.parquet("s3://out")
While the code looks deceptively simple, Spark does a lot of work under the hood, such as grouping the data into Parquet partitions, ensuring each record appears in the output exactly once, and recovering the job’s state if you restart it. Finally, to serve this data interactively instead of writing it to Parquet, we could just change writeStream to use the (currently alpha) in-memory sink and connect a JDBC client to Spark to query it.
Our long-term vision for streaming in Spark is ambitious: we want every library in Spark to work in an incremental fashion on Structured Streaming. Although this is a big goal, Apache Spark is well positioned to achieve it. Its libraries are already built on common, narrow APIs (RDDs and DataFrames), and Structured Streaming is designed explicitly to give results consistent with these unified interfaces.
The biggest insight in Spark since its beginning is that developers need unified interfaces. For example, batch computation on clusters used to require many disjoint systems (MapReduce for ETL, Hive for SQL, Giraph for graphs, etc), complicating both development and operations. Spark unified these workloads on one engine, greatly simplifying both tasks. The same insight applies to streaming. Because streaming workloads are usually part of a much larger continuous application, which may include serving, storage, and batch jobs, we want to offer a unified API and system for building end-to-end continuous applications.
Our Structured Streaming model blog post explores the streaming API and execution model in more detail. We recommend you read this post to get started with Structured Streaming.
In addition, the following resources cover Structured Streaming:
- Spark 2.0 and Structured Streaming
- Structuring Spark: DataFrames, Datasets and Streaming
- A Deep Dive Into Structured Streaming
- Structured Streaming Programming Guide | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100508.42/warc/CC-MAIN-20231203161435-20231203191435-00500.warc.gz | CC-MAIN-2023-50 | 9,463 | 33 |
http://msdn.microsoft.com/en-us/library/system.windows.controls.textbox.shouldserializetext.aspx | code | Returns a value that indicates whether the effective value of the Text property should be serialized during serialization of the TextBox object.
public bool ShouldSerializeText(
A serialization service manager object for this object.
manager is null.
Windows 8.1, Windows Server 2012 R2, Windows 8, Windows Server 2012, Windows 7, Windows Vista SP2, Windows Server 2008 (Server Core Role not supported), Windows Server 2008 R2 (Server Core Role supported with SP1 or later; Itanium not supported)
The .NET Framework does not support all versions of every platform. For a list of the supported versions, see .NET Framework System Requirements. | s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1410657137841.57/warc/CC-MAIN-20140914011217-00091-ip-10-234-18-248.ec2.internal.warc.gz | CC-MAIN-2014-41 | 642 | 6 |
https://www.experts-exchange.com/questions/24997616/DNS-consiguration-question-s.html | code | I have added a new site into my Server 03 AD Domain. Right now there is only one DC in this new site. At this point, this server is providing AD DC, DNS, WINS. DHCP is being handeled by another device in a seperate Domain I am sharing this "Space" with (for now).
I have two other sites: the main site, and a smaller single DC site.
This new site has a pair of linked Riverbed devices between the new site, and the main site.
What I am trying to figure out is:
Network neighborhood from the new DC is SLOW to show me anything. When I look at the list of machines in the domain, this new DC does not show up. I have compared my DNS setup between all three sites, and the only two differences I see are:
The "Other Two" sites have the same set of two external Forwarders in the DNS settings. I believe these are the ISP DNS servers for the main office. The forwarders on the new DC are the external DNS servers from the ISP for that office.
And, Each of the other two sites DC's have an SOA record that basically points to itself, where as the new DC's SOA record points to the main DC server from the main office.
...What am I missing here? | s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046155458.35/warc/CC-MAIN-20210805063730-20210805093730-00140.warc.gz | CC-MAIN-2021-31 | 1,139 | 8 |
https://live.yworks.com/demos/layout-features/tree/ | code | Basic Tree Layout
This demo shows basic configuration options for the
The global layout orientation is changed
to left-to-right to get a horizontal tree layout.
A DefaultNodePlacer is specified as placer for the tree nodes, its
distance settings are configured to get a rather wide spacing between elements.
The port assignment is configured
such that the edges are distributed at the node border and the segments are forced to
keep some distance in this example (via
minimumChannelSegmentDistance). By default, the tree layout groups the edge segments and places the ports in the
Edges from the parent to the child nodes are ordered with respect to the label text of
the child nodes by defining an appropriate
edge comparer function.
You can copy the code snippet to configure the layout from
The Developer's Guide provides more in-depth information about the
Tree Layout provided by yFiles. | s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817674.12/warc/CC-MAIN-20240420184033-20240420214033-00554.warc.gz | CC-MAIN-2024-18 | 892 | 16 |
https://zenodo.org/record/6123879 | code | Dataset Open Access
Four multimedia recommender systems datasets to study popularity bias and fairness:
Each dataset contains of user interactions (user_events.txt) and three user groups that differ in their inclination to popular/mainstream items: LowPop (low_main_users.txt), MedPop (med_main_users.txt), and HighPop (high_main_users.txt).
The format of the three user files are "user,mainstreaminess"
The format of the user-events files are "user,item,preference"
Example Python-code for analyzing the datasets as well as more information on the user groups can be found on Github (https://github.com/domkowald/FairRecSys) and on Arxiv (https://arxiv.org/abs/2203.00376) | s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233506480.7/warc/CC-MAIN-20230923094750-20230923124750-00790.warc.gz | CC-MAIN-2023-40 | 673 | 6 |
https://helpdeskgeek.com/how-to/disable-snap-assist-in-windows-7-8-10/ | code | Because some find it very annoying
Starting with Windows 7, Microsoft added Aero Snap, a part of the Aero Desktop experience not found in Windows Vista’s Aero Desktop. With Aero Snap, users can drag any window to the top, left, or right side of the screen to maximize the window in a special way.
Dragging a window to the top maximizes the window to full screen while dragging it either left or right maximizes the window to half screen depending on the side to which you drag it. Although useful for when you want to see two windows at the same time, some users find that the Aero Snap feature is a bit too aggressive, maximizing windows when the user did not intend to.
Luckily, Microsoft lets you stop the Aero Snap feature from operating without having to disable Aero Desktop altogether or making some complicated Registry entry.
Disable the Aero Snap Feature in Windows 7/8/10
To disable the Aero Snap feature, begin by clicking on Start and then entering Ease of Access Center into the Search Programs and Files search box. Press Enter.
Not surprisingly, you should now be looking at the Ease of Access Center, which is a central location for all ease of access options available in Windows 7/8/10. At the bottom of the list is an option titled Make it Easier to Focus on Tasks. Click on this option and you will be looking at the Make it Easier to Focus on Tasks window.
Near the bottom of the window, you will see an option titled Make it Easier to Manage Windows. Under this option, check the box titled Prevent Windows from Being Automatically Arranged When Moved to the Edge of the Screen.
Oddly, Microsoft didn’t just call this option Disable Aero Snap like it probably should have. Doing so would have made finding and disabling this feature a whole lot easier.
Why Disable Windows 7/8/10 Aero Snap?
Although a seemingly useful feature, some users find the implementation of Aero Snap more of an annoyance than a helpful feature. Users of previous versions of Windows could arrange two windows side by side by right clicking on the taskbar and choosing a menu option. Unfortunately, this method was frustrating when users had more than two windows minimized to the taskbar.
If you often have multiple applications open such as Word and Excel while you work, you probably spend a lot of time moving windows around to see what going on. Often, this includes moving a window partially off the desktop to view what’s underneath.
This is usually faster than minimizing the window to the taskbar for just a few seconds before having to restore the window again. With Aero Snap enabled, many users find that this feature actually slows them down when they accidently move a window too far to the top, left, or right of the screen. Enjoy! | s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947473824.45/warc/CC-MAIN-20240222193722-20240222223722-00514.warc.gz | CC-MAIN-2024-10 | 2,750 | 13 |
https://db0nus869y26v.cloudfront.net/en/Progressive_Graphics_File | code | |Internet media type||image/x-pgf, application/octet-stream|
|Developed by||xeraina GmbH|
|Type of format||wavelet-based bitmapped image format|
|Extended from||JPEG, PNG|
PGF (Progressive Graphics File) is a wavelet-based bitmapped image format that employs lossless and lossy data compression. PGF was created to improve upon and replace the JPEG format. It was developed at the same time as JPEG 2000 but with a focus on speed over compression ratio.
PGF can operate at higher compression ratios without taking more encoding/decoding time and without generating the characteristic "blocky and blurry" artifacts of the original DCT-based JPEG standard. It also allows more sophisticated progressive downloads.
PGF supports a wide variety of color models:"PGF Facts Sheet" (PDF).
PGF claims to achieve an improved compression quality over JPEG adding or improving features such as scalability. Its compression performance is similar to the original JPEG standard. Very low and very high compression rates (including lossless compression) are also supported in PGF. The ability of the design to handle a very large range of effective bit rates is one of the strengths of PGF. For example, to reduce the number of bits for a picture below a certain amount, the advisable thing to do with the first JPEG standard is to reduce the resolution of the input image before encoding it — something that is ordinarily not necessary for that purpose when using PGF because of its wavelet scalability properties.
The PGF process chain contains the following four steps:
Initially, images have to be transformed from the RGB color space to another color space, leading to three components that are handled separately. PGF uses a fully reversible modified YUV color transform. The transformation matrices are:
The chrominance components can be, but do not necessarily have to be, down-scaled in resolution.
The color components are then wavelet transformed to an arbitrary depth. In contrast to JPEG 1992 which uses an 8x8 block-size discrete cosine transform, PGF uses one reversible wavelet transform: a rounded version of the biorthogonal CDF 5/3 wavelet transform. This wavelet filter bank is exactly the same as the reversible wavelet used in JPEG 2000. It uses only integer coefficients, so the output does not require rounding (quantization) and so it does not introduce any quantization noise.
After the wavelet transform, the coefficients are scalar-quantized to reduce the amount of bits to represent them, at the expense of a loss of quality. The output is a set of integer numbers which have to be encoded bit-by-bit. The parameter that can be changed to set the final quality is the quantization step: the greater the step, the greater is the compression and the loss of quality. With a quantization step that equals 1, no quantization is performed (it is used in lossless compression). In contrast to JPEG 2000, PGF uses only powers of two, therefore the parameter value i represents a quantization step of 2i. Just using powers of two makes no need of integer multiplication and division operations.
The result of the previous process is a collection of sub-bands which represent several approximation scales. A sub-band is a set of coefficients — integer numbers which represent aspects of the image associated with a certain frequency range as well as a spatial area of the image.
The quantized sub-bands are split further into blocks, rectangular regions in the wavelet domain. They are typically selected in a way that the coefficients within them across the sub-bands form approximately spatial blocks in the (reconstructed) image domain and collected in a fixed size macroblock.
The encoder has to encode the bits of all quantized coefficients of a macroblock, starting with the most significant bits and progressing to less significant bits. In this encoding process, each bit-plane of the macroblock gets encoded in two so-called coding passes, first encoding bits of significant coefficients, then refinement bits of significant coefficients. Clearly, in lossless mode all bit-planes have to be encoded, and no bit-planes can be dropped.
Only significant coefficients are compressed with an adaptive run-length/Rice (RLR) coder, because they contain long runs of zeros. The RLR coder with parameter k (logarithmic length of a run of zeros) is also known as the elementary Golomb code of order 2k.
There are several self-proclaimed advantages of PGF over the ordinary JPEG standard:
The author published libPGF via a SourceForge, under the GNU Lesser General Public License version 2.0. Xeraina offers a free Photoshop
.8bi file format plugin, a Win32 console encoder and decoder, and PGF viewers based on WIC for 32bit and 64bit Windows platforms. Other WIC applications including Photo Gallery are able to display PGF images after installing this viewer.
Digikam is a popular open-source image editing and cataloging software that uses libPGF for its thumbnails. It makes use of the progressive decoding feature of PGF images to store a single version of each thumbnail, which can then be decoded to different resolutions without loss, thus allowing users to dynamically change the size of the thumbnails without having to recalculate them again.
.pgf and the TLA PGF are also used for unrelated purposes:
.pgfwith Portfolio Graphics.
((cite web)): External link in | s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296948900.50/warc/CC-MAIN-20230328232645-20230329022645-00702.warc.gz | CC-MAIN-2023-14 | 5,380 | 24 |
https://www.youtube.com/howden-video | code | You may or may not have heard about our SmartEXEC™ technology. Here we gather a collection of videos to showcase our Ventilation on Demand solution. This automation and optimisation tool offers great benefits to the mining industry. Subscribe to our channel: https://goo.gl/Tt07H2
Howden has been setting standards in compressor engineering for more than 160 years. We are the originators of diaphragm, twin screw and Roots® technologies, and have introduced major improvements in centrifugal and reciprocating compressors. | s3://commoncrawl/crawl-data/CC-MAIN-2017-39/segments/1505818687447.54/warc/CC-MAIN-20170920194628-20170920214628-00459.warc.gz | CC-MAIN-2017-39 | 526 | 2 |
https://secretgeek.net/Category/movie | code | (See also: all categories or the complete list of articles)
NimbleText is now a Machine Learning Platform
Using Computers to Invent New and *Fun* European Style Board Games
Have you ever seen the International Space Station?
Launching the secretGeek Wiki
Go and read a book.
Do *NOT* try this Hacking Script at home
Harvey, a .net chat server built with RabbitMQ
[code], color, commandline, css, functional, html, it industry, microsoft, movie, rabbitmq, security, sql, tools, UX
On Task Hoarding and Todo Bankruptcy
blog, google, html, linguistics, microISV, movie, nimbletext, productivity, [ruby], sql, tools
The Correct Order for a First Time Viewing of The Lord Of The Rings
I went to write down a book name and founded an internet empire instead.
amazon, apple, blog, facebook, food, games, movie, social network, tools, You-Must-Get-It
I see you're using vim. Let me fix that for you.
The Movie Hollywood (And My Wife) Doesn't Want You To See: Weekend at Jacko's
Movie: Priest Academy
Original Premise for a road movie
blog, legal, movie, UX, working blue
10 Simple Rules To Follow In Case Your Software Becomes Self-Aware
A truly original thought
TimeSnapper 3.0 -- an interactive, bubbling cauldron of possibilities
Workflow software: I'm calling the bluff.
More about the wiki adventure game...
TimeSnapper 2.7: Word Clouds reveal your computing lifestyle!
52 great reasons to fail at software
atwood, blog, [code], color, food, html, it industry, microISV, movie, working blue
TimeSnapper Special - 9 More Days
How to block websites that kill your productivity.
blog, [code], facebook, html, microISV, microsoft, movie, podcast, productivity, robots, social network, tools
Don't Forget The Caret^ and the $tick
The Principle of Scale: A fundamental lesson they failed to teach us at school
MicroISV: Step 4 of 25 -- Basic Website Content
agile, blog, editor, education, html, linguistics, microISV, movie, sql, sysadmin, tdd, tools, UX, working blue
Top Tips of 2007, part I: Leon's One-Step Guide to Making a Nature Appreciation Film in Africa
Gunderloy, Spolsky, Atwood: three big idiots.
atwood, blog, cartoon, education, hanselman, html, ideas rat, microsoft, movie, ocr, productivity, [ruby], UX
Unlikely to fill Gmail account before I die
TimeSnapper 2 Goes Live!
Gripping Reading, All The Way Through!
TimeSnapper 2.0 is near to completion
Forget Intellisense: I want IntelliSEARCH
FU: Findability precedes Usability.
Lessons from Amazon referrals
amazon, coffee, commandline, education, food, functional, google, html, it industry, movie
Pay Day for the Ideas Rat
apple, australia, blog, css, editor, google, hanselman, html, ideas rat, microsoft, movie, tools, UX
Save Dog as Virtual Chicken
A Gmail Wiki.
Just keep zooming -- the resolution never degrades.
Email Personality Quiz
Anti-Trust: bad geek movie
(See also: All Years, all categories or the complete list of articles) | s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296943483.86/warc/CC-MAIN-20230320114206-20230320144206-00411.warc.gz | CC-MAIN-2023-14 | 2,898 | 53 |
https://www.geeksforgeeks.org/activity-lifecycle-in-android-with-demo-app/ | code | Activity Lifecycle in Android with Demo App
In Android, an activity is referred to as one screen in an application. It is very similar to a single window of any desktop application. An Android app consists of one or more screens or activities.
Each activity goes through various stages or a lifecycle and is managed by activity stacks. So when a new activity starts, the previous one always remains below it. There are four stages of an activity.
- If an activity is in the foreground of the screen i.e at the top of the stack, then it is said to be active or running. This is usually the activity that the user is currently interacting with.
- If an activity has lost focus and a non-full-sized or transparent activity has focused on top of your activity. In such a case either another activity has a higher position in multi-window mode or the activity itself is not focusable in the current window mode. Such activity is completely alive.
- If an activity is completely hidden by another activity, it is stopped or hidden. It still retains all the information, and as its window is hidden thus it will often be killed by the system when memory is needed elsewhere.
- The system can destroy the activity from memory by either asking it to finish or simply killing its process. When it is displayed again to the user, it must be completely restarted and restored to its previous state.
For each stage, android provides us with a set of 7 methods that have their own significance for each stage in the life cycle. The image shows a path of migration whenever an app switches from one state to another. | s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882573399.40/warc/CC-MAIN-20220818185216-20220818215216-00289.warc.gz | CC-MAIN-2022-33 | 1,601 | 8 |
https://www.fiverr.com/lees6620/transcribe-3-minutes-of-cantonese-audio | code | I am fast, accurate and reliable. Send me any file or URL and I will transcribe the audio in Cantonese (up to 2 minutes) into a word document for $5.
I will transcribe anything you send me be it interview, video, webinar, podcast, telephone, dictation or radio. I will transcribe any online content like YouTube for better SEO in Google etc.
By default, I will transcribe it into written Traditional Chinese (it's a common practice so all Chinese speakers can understand). If you need me to transcribe it into written Cantonese (non-Cantonese Chinese speakers will have difficulties in understanding), please let me know in advance. | s3://commoncrawl/crawl-data/CC-MAIN-2016-40/segments/1474738661213.28/warc/CC-MAIN-20160924173741-00212-ip-10-143-35-109.ec2.internal.warc.gz | CC-MAIN-2016-40 | 632 | 3 |
https://forums.afterdawn.com/threads/need-help-with-ut2004-on-dvd-backup-please.308920/ | code | hi, I have a legit copy of UT2004 but its on DVD and I want to back it up, but I don't have a dvd burner, so I wanted to back it up to cd. Well, when I copied all the files from the dvd to my hard drive, it actually was already broken down into cd1, cd2, etc. So I started to copy each one to a disk. But when I went to test it out and reinstall the game using my newly made cds, it was saying something about missing files, and if I was installing off a cd to try to clean the cd and retry. Well, I thought it might be doing this because it was on a dvd which is one disk, and now its on 6 cds, so I put in the second cd to see if it would pick up the installation where it stopped but it did not. So my question is this, does anyone know how to make it to where it knows to ask for the second cd and so on and so forth as to make a good backup on cds? Thanks in advance! | s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257649095.35/warc/CC-MAIN-20180323220107-20180324000107-00009.warc.gz | CC-MAIN-2018-13 | 872 | 1 |
https://alternativeto.net/software/youtube-go/about/ | code | What is YouTube Go?
YouTube Go provides a video experience that focuses on offline support, letting you download and share video content locally. Preview videos when offline and share them with nearby friends and download videos to watch later.
- Download and watch your favorite videos
- Instantly share videos - no data used to transfer
- Search for any video and discover new videos
- Only 9MB app size
Google announced that Youtube Go will shut down in August 2022: https://support.google.com/youtube/thread/162222567/youtube-go-is-going-away-in-august-of-this-year
YouTube Go Screenshots
YouTube Go information | s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764500288.69/warc/CC-MAIN-20230205193202-20230205223202-00623.warc.gz | CC-MAIN-2023-06 | 615 | 9 |
https://methodsblog.com/2015/02/19/issue-6-2/ | code | Issue 6.2 is now online!
The February issue of Methods is now online!
This month we have two applications articles. Both are free to access, no subscription required.
– NLMpy: A PYTHON software package for the creation of neutral landscape models (there are also two videos associated to this paper on our Youtube channel)
– BAT – an R package for the measurement and estimation of alpha and beta taxon, phylogenetic and functional diversity
There are also two OnlineOpen articles in this month’s issue. Power analysis for generalized linear mixed models in ecology and evolution, by Paul C. D. Johnson,Sarah J. E. Barry, Heather M. Ferguson and Pie Müller, focuses on why and how we use power analysis for GLMMs using simulations more than we should.
Our second Open Access article is also the source of our cover image, which shows the ciliate protist Paramecium caudatum (about 0.25 mm long). Protist species like this are commonly found in aquatic habitats and offer a unique study system to test ecological and evolutionary concepts. The protist was isolated from a natural pond and subsequently used for microcosm experiments, which have a long tradition in order to test ecological and evolutionary concepts.
In the accompanying review paper, Florian Altermatt et al. describe a wide range of available techniques to use this and many other protists species to conduct microcosm experiments. The review paper gives detailed protocols of available techniques with a focus on modern, high-frequency and high-throughput measurements, and outlines how such microcosm experiments may be used to address a wide range of questions.
This comprehensive guide to using protist microcosms as a model system in ecology and evolution in ‘Big answers from small worlds‘, which is available free of charge.
Photo © Regula Illi and Florian Altermatt.
To keep up to date with Methods newest content, have a look at our Accepted Articles and Early View articles, which will be included in forthcoming issues. | s3://commoncrawl/crawl-data/CC-MAIN-2023-23/segments/1685224649105.40/warc/CC-MAIN-20230603032950-20230603062950-00567.warc.gz | CC-MAIN-2023-23 | 2,011 | 11 |
https://gis.stackexchange.com/questions/44410/splitting-feature-class-by-unique-values-for-an-attribute-field-in-arcgis-for-de?noredirect=1 | code | I would like to split a single shapefile into multiple shapefiles by using unique values for one attribute/field to select features for each value out into a new feature class.
I found ET Geowizard to do this but it's not free.
In arcscripts.esri.com I found a tool named "SplitLayerByAttributes" but it does not work in ArcMap . (see comments below, this was an old version which has been updated. The new version can be found at resources.arcgis.com )
Do you know of any tools or solution to split by attribute in the way described above? | s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178363072.47/warc/CC-MAIN-20210301212939-20210302002939-00477.warc.gz | CC-MAIN-2021-10 | 540 | 4 |
https://math.stackexchange.com/questions/3350215/definition-of-entropy-information-theory | code | In Information Theory, entropy is defined as:
where $-P_ilog(P_i)$ looks like this (using log base 2):
From just a generic English definition of entropy, meaning lack of predictability, I don't find this particularly intuitive. Would you not have the most entropy (be the most uncertain) when $P_i=0.5$? In other words, would it not make more sense to use a measure like this:
where $-4P_i(P_i-1)$ looks like this:
instead? What is the advantage of using $P_ilog(P_i)$? | s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496665575.34/warc/CC-MAIN-20191112151954-20191112175954-00011.warc.gz | CC-MAIN-2019-47 | 469 | 5 |
http://crypto.stackexchange.com/tags/multiple-encryption/new | code | New answers tagged multiple-encryption
If there is a logger malware in the computer, not a keylogger, the cascading cipher should be weakened, because the application will know the specific way the cipher encrypted something. Reverse mathematic operations maybe discover the password. Cascading ciphers are secure, and have chances that not even a Quantum computer will break it. Here i only say ...
Top 50 recent answers are included | s3://commoncrawl/crawl-data/CC-MAIN-2016-18/segments/1461860122501.26/warc/CC-MAIN-20160428161522-00080-ip-10-239-7-51.ec2.internal.warc.gz | CC-MAIN-2016-18 | 434 | 3 |
http://al0sh11.blogspot.com/ | code | جديد العاب فلاش 2013 - 2014
A young Ninja must survive a deadly training out in the woods and uncover the Secret Ninja scrolls in order to learn how the master the art of the Shadow Warrior. This game plays unfair as all the obstacles are hidden and you really need to use your ninja sense in order to avoid getting trapped and fullfill your destiny.
Eliminate all of the opposing tanks in each match.
- Select a weapon from the bottom of the screen when it's your turn.
- Use your mouse to aim, click to fire.
- You can move by clicking the 'Move' button and clicking where you want your tank to relocate (short distance).
- You can either fire OR move during each turn.
- You have 15 seconds | s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547584519757.94/warc/CC-MAIN-20190124080411-20190124102411-00435.warc.gz | CC-MAIN-2019-04 | 706 | 8 |
https://kolibri-dev.readthedocs.io/en/develop/development_workflow.html | code | At a high level, we follow the ‘Gitflow’ model. Some helpful references:
In particular, please use the labels “Needs review”, “Work in progress”, and “Needs updates” mutually exclusively to communicate the state of the PR.
Every pull request will require some combination of manual testing, code review, automated tests, gherkin stories, and UI design review. Developers must fully test their own code before requesting a review, and then closely follow the template and checklist that appears in the PR description. All automated tests must pass.
Unit tests and gherkin stories should be written to ensure coverage of critical, brittle, complicated, or otherwise risky paths through the code and user experience. Intentional, thoughtful coverage of these critical paths is more important than global percentage of code covered.
Try to keep PRs as self-contained as possible. The bigger the PR, the more challenging it is to review, and the more likely that merging will be blocked by various issues. If your PR is not being reviewed in a timely manner, reach out to stakeholders and politely remind them that you’re waiting for a review.
Some additional guidelines:
Submitters should fully test their code before asking for a review
If the PR is languishing, feel free to prod team members for review
Try to keep the PR up-to-date with the target branch
Make sure to use the checkboxes in the PR template
Within the Kolibri repo, we have the following primary rule:
Never rewrite history on shared branches.
History has been rewritten if a force push is required to update the remote. This will occur from e.g. amending commits, squashing commits, and rebasing a branch.
Some additional git history guidance:
Be encouraged to rewrite history on personal branches so that your git commits tell a story
Avoid noisy, meaningless commits such as “fixed typo”. Squash these prior to submitting a PR
When possible, make each commit a self-contained change that plays nicely with
Once a PR code review has occurred, avoid squashing subsequent changes as this makes it impossible to see what changes were made since the code review
Don’t worry too much about a “clean” commit history. It’s better to have some messy commits than to waste an hour than debugging a rebase-gone-wrong
When reviewing PRs, keep feedback focused on critical changes. Lengthy conversations should be moved to a real-time chat when possible. Be polite, respectful, and constructive. We highly recommend following the guidance in this blog post.
Some general guidelines:
Reviewers should actually run and test the PR
When giving opinions, clarify whether the comment is meant to be a “blocking” comment or if it is just a conversation
Pre-existing issues or other cleanup suggestions are can be opened as new issues, or mentioned as “non-blocking” comments
Code formatting comments should be rare because we use Prettier and Black
Finally, if you see a very trivial but important necessary change, the reviewer can commit the change directly to a pull request branch. This can greatly speed up the process of getting a PR merged. Pushing commits to a submitter’s branch should only be done for non-controversial changes or with the submitter’s permission.
When pushing to another user’s branch, you may get an error like:
Authentication required: You must have push access to verify locks
Remember to keep the “Needs review”, “Work in progress”, and “Needs updates” mutually exclusive and up-to-date.
Who should merge PRs, and when?
First, all automated checks need to pass before merging. Then…
If there is just one reviewer and they approve the changes, the reviewer should merge the PR immediately
If there are multiple reviewers or stakeholders, the last one to approve can merge
The reviewer might approve the PR, but also request minor changes such as a typo fix or variable name update. The submitter can then make the change and merge it themselves, with the understanding that the new changes will be limited in scope
Stale reviews should be dismissed by the PR submitter when the feedback has been addressed
We have the following release types:
Info: major, minor, patch
v1.2.3on a release branch
Final integration testing, string freeze, and beta release candidates
High level of risk-aversion in PRs
Info: major, minor, patch, beta
v1.2.3-beta4on a release branch
Initial testing releases
Avoid broken builds in PRs
Info: major, minor, patch, alpha
v1.2.3-alpha4on the develop branch
Feature branches, PRs, or other git commits
Info: major, minor, patch, commit
Experimental work is OK
Within the Learning Equality Kolibri repository:
developbranch is our current development branch, and the default target for PRs
Release branches named like
release-v1.2.x(for example). This will track all patch releases within the 1.2.x minor release line. Distinct releases are tracked as tags like
We sometimes create feature branches for changes that are long-running, collaborative, and disruptive. These should be kept up-to-date with
developby merging, not rebasing.
If a change needs to be introduced to an older release, target the oldest release branch that we want the change made in. Then that change will need to be merged into all subsequent releases, one-at-a-time, until it eventually gets back to
We use a wide range of labels to help organize issues and pull requests in the Kolibri repo.
These are used to sort issues and sometimes PRs by priority if and only if the item is assigned a milestone. Every issue in a milestone ought to have a priority label.
Only ‘critical’ items are strictly blockers for a release, but typically all important items should be expected to make it in, too. Priority within a release is generally assigned by a core Learning Equality team member.
P0 - critical
P1 - important
P2 - normal
P3 - low
The changelog label is used on PRs or issues to generate ‘more details’ links in the Release Notes.
The work-in-progress label is helpful if you have a PR open that’s not ready for review yet.
Labels prefixed with DEV: are used to help organize issues (and sometimes PRs) by area of responsibility or scope of domain knowledge necessary.
Labels prefixed with TODO: help flag items that need some action before the issue or PR can be fully resolved. | s3://commoncrawl/crawl-data/CC-MAIN-2021-43/segments/1634323585518.54/warc/CC-MAIN-20211022181017-20211022211017-00225.warc.gz | CC-MAIN-2021-43 | 6,327 | 67 |
https://www.aminer.cn/pub/5a73cbcc17c44a0b3035f7b3/mastering-chess-and-shogi-by-self-play-with-a-general-reinforcement-learning | code | Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm
arXiv: Artificial Intelligence, Volume abs/1712.01815, 2017.
The game of chess is the most widely-studied domain in the history of artificial intelligence. The strongest programs are based on a combination of sophisticated search techniques, domain-specific adaptations, and handcrafted evaluation functions that have been refined by human experts over several decades. In contrast, the AlphaGo Zero p...More
PPT (Upload PPT) | s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610703533863.67/warc/CC-MAIN-20210123032629-20210123062629-00041.warc.gz | CC-MAIN-2021-04 | 512 | 4 |
https://seanoreilly.org/2019/01/11/emacs-min.html | code | TLDR; a minimalist config file to make vanilla Emacs work as expected on macOS.
Times change. Needs change.
Like many, I’ve spent a lot of time adding tweaks to my Emacs config files to make sure that common tasks and actions are only one key stroke away, and that specific packages behave just so. I’ve spent hours making sure that my config is well documented and structured in something like an org file. This can lead to a sense of attachment with one’s configuration files (be it in Emacs or any other tool).
However, every now and then, I realise that I have way too much going on in my Emacs setup. This is largely due to me doing a bad job of removing packages that are no longer relevant to my current workflow.
Bloated config files generally lead to slow Emacs initialisation times. This can send me into an irrational frenzy where I believe the only sane thing to do is to start over by trashing my current configuration. I’ve found that the best thing to do in this situation is to breathe deeply, calm down, and then delete my configuration files.
Having all my config in a single file, whether it is in org-mode or Emacs Lisp format does not scale well, and creates friction with respect to future maintainability. The plan going forward is to split related package configurations into separate Emacs Lisp files that I can load independently with a single line of code each. But that is a blog post for another time. For now I just want vanilla Emacs.
I’ve found that there are a few rough edges with how vanilla Emacs (v26.1 at time of writing) behaves on macOS. This generally leads me to the same StackOverflow articles I looked at the last time I went through this kind of spring clean. What I need is just enough setup to make vanilla Emacs work as expected on macOS.
The minimalist setup
This is what I have so far.
;; ~/.emacs.d/init.el ;; A minimalist setup for Emacs on macOS ;; Basically just fixes stuff that is misconfigured by default ;; ui (tool-bar-mode -1) (scroll-bar-mode -1) ;; fix env vars on macOS (setenv "PATH" (concat (getenv "PATH") ":/usr/local/bin")) ;; often GPG decryption simply freaks out - this enforces the password prompt in the mini buffer (setq epa-pinentry-mode 'loopback)
As you can see, that’s not very much code. That’s kind of what I am going for. | s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912201329.40/warc/CC-MAIN-20190318132220-20190318154220-00111.warc.gz | CC-MAIN-2019-13 | 2,316 | 11 |
https://harbormaster.ai/rails-tech-stack/ | code | Stack Name - Ruby on Rails
While Harbormaster will handle creating everything required of CRUD functionality for each entity in a model, this version of the Rails tech stack does not yet handle generating everything for associations between entities.
|Long Name||Base Rails Tech Stack|
Generated Project Support:
The core templates of this tech stack are as follows:
- Vendor specific handlers using Velocity template files and macros to generate a pipeline YAML file for CI functionality.
- Create a Ruby model for each entity found in the entity model. Each corresponds to a table in the database. Attributes are not declared here, but are instead declared in the initial migration schema. However, Enumerated types are explicitly declared here.
- Creates a form and table view for each entity in the entity model.
- The main and sub layouts for the application
- Used to handle URL invocation with interaction to Rails underlying system that handles the model lifecycle
- For each entity, all appropriate routes for related CRUD actions
- An initial migration schema file to get the database schema created for the provided entity model
- For each entity and its attributes, creates a single file with defaults data values to create multiple instances. This is useful for testing
- For each entity, creates a test for the associated controller to create, update, and delete that entity type. | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100568.68/warc/CC-MAIN-20231205204654-20231205234654-00247.warc.gz | CC-MAIN-2023-50 | 1,394 | 14 |
https://bridgingbarriers.utexas.edu/news/good-systems-awards-summer-fellowships-faculty-researchers-and-graduate-students | code | Good Systems has awarded 20 summer fellowships to faculty, researchers, and graduate students from 13 departments and units across UT Austin to advance the mission of this research grand challenge by funding innovative research that investigates the ethical implications of AI technologies in society.
Fellows will conduct their proposed research projects from June to August, culminating in presentations of their findings in Fall 2022.
Faculty and Researcher Fellowship Awardees:
- John Robert Bautista, Postdoctoral Fellow, School of Information, “Development and psychometric validation of the AI Ethical Concerns Inventory - Healthcare (AIECIH)”
- Avinash Collis, Assistant Professor, Department of Information, Risk, & Operations Management, “AI and privacy: How knowledge about AI technologies impacts peoples’ valuations of personal data”
- Arya Farahi, Assistant Professor, Department of Statistics and Data Sciences, “A computational framework for modeling structural disparity”
- Daniel Koehler, Assistant Professor, School of Architecture, “Communal physics: Building synthetic data sets or how to learn from future architectures?”
- Yan Leng, Assistant Professor, Department of Information, Risk, & Operations Management, “The exposure risk of social interactions when online behavioral data meets artificial intelligence”
- Erin McElroy, Assistant Professor, Department of American Studies, “Landlord technology and tenant harms”
- Vagheesh Narasimhan, Assistant Professor, Department of Integrative Biology, “Good AI systems for ethical and transferrable genomic risk prediction”
- Jaganath Sankaran, Assistant Professor, Lyndon B Johnson School of Public Affairs, “Machine learning in nuclear missile hunting: Possibilities and perils”
- Ciaran Trace, Associate Professor, School of Information, and James Hodges, Postdoctoral Fellow, School of Information, “Information governance and algorithmic accountability”
- Junmin Wang, Professor, Department of Mechanical Engineering, “Study on ethical and socially-compliant AI-powered driving behaviors for autonomous driving”
Graduate Student Fellowship Awardees:
- Liz Bender, School of Information and Department of English, “Moral machines? Determining ethics using rhetorical analysis of AI- vs. human-generated news articles”
- Mehar Jaiswal, Department of Aerospace Engineering and Engineering Mechanics, “Autonomous one-wheel home furniture robot: Bumbybot”
- Jared Jensen, Department of Communication Studies, “Investigating how AI transforms creative collaboration in music”
- Eunhye Ko, Department of Curriculum and Instruction, “Value sensitive design to integrate XAI in education: Empowering students with agency, acceptance, and trust”
- Sheila Lalwani, School of Journalism and Media, “Digital guardians? How data protection officers (do and don’t) navigate through ethical dilemmas: Important lessons for the U.S.”
- Hakeoung Hannah Lee, Department of Curriculum and Instruction, “Responsible AI: Why did mentors hide from mentees they had seen their mentees’ learning analytics data?”
- Sooyong Lee, Department of Computer Science, “Rationale-based active learning for BERT-to-BERT”
- Jonathan Malott, School of Architecture, “AI-generated architecture and building generation as a tool for novel design”
- Nandini Sharma, Department of Communication Studies, “Space, time, and coding practices: A coder's tale of AI ethics”
- Yuqing Wang, School of Architecture, “UX technical framework of smart library based on metaverse”
Faculty, research scientists, and postdoctoral fellows will be awarded $4,000 per fellowship, while master’s and doctoral students will be awarded $6,000.
Applications were reviewed by the Good Systems Executive Team and evaluated based on creativity, feasibility, engagement, subject matter expertise, and evidence of prior research accomplishments.
One of three Bridging Barriers research grand challenges, Good Systems brings together faculty, researchers, and students in campus-wide, interdisciplinary teams to investigate how to define, evaluate, and build socially beneficial AI-based technologies – or “good systems.”
Good Systems is a research grand challenge at The University of Texas at Austin. We’re a team of information and computer scientists, robotics experts, engineers, humanists and philosophers, policy and communication scholars, architects, and designers. Our goal over the next eight years is to design AI technologies that benefit society. | s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233506623.27/warc/CC-MAIN-20230924055210-20230924085210-00121.warc.gz | CC-MAIN-2023-40 | 4,564 | 28 |
https://medium.com/@breich_84283/enabling-network-discovery-on-vista-without-windows-server-2008-f02319e66d93 | code | Note: This article was originally written on reich-consulting.net. Reich Web Consulting has narrowed its focus to the web and no longer offers tech support services, so we’ve moved all of our tech support content off-site. We hope you find this article useful. It is provided as-is, and we will no longer provide support on this topic.
This problem has been making my life a living hell since we deployed Windows Vista to a handful of workstations several months ago. According to Microsoft, “Network discovery is a network setting that affects whether your computer can see (find) other computers and devices on the network and whether other computers on the network can see your computer.” What this means to a network administrator is that, even with correct firewall rules in place, communication between your Vista workstations and 2000/2003/XP workstations and servers will basically cease to function. Apparently Windows Server 2008 has a GPO policy that resolves this, but the fact that it hasn’t been released to the public yet tends to keep up from upgrading. In the meantime you can solve this problem by applying a couple of registry entries to your workstations.
[Update]: For those of you fortunate enough to be using Server 2008, check out this page over at the MSDN Blogs about turning on Network Discovery via Server 2008 Group Policy.
Windows Registry Editor Version 5.00[HKEY_LOCAL_MACHINE\SOFTWARE\Policies\Microsoft\Windows\LLTD] | s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655897027.14/warc/CC-MAIN-20200708124912-20200708154912-00048.warc.gz | CC-MAIN-2020-29 | 1,458 | 4 |
http://stackoverflow.com/questions/13667515/embed-an-external-link-in-django-app/13667561 | code | I added a link in myproject/templates/items/index.html like this
<a href="www.google.com">Search More</a>
My hope was that when a user clicks it, she would visit google.com. However, the result right now is this link is "myhost/items/www.google.com" instead of "www.google.com".
(I am modifying an existing app, and I have to admit I have not fully read django doc yet) | s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1408500804220.17/warc/CC-MAIN-20140820021324-00236-ip-10-180-136-8.ec2.internal.warc.gz | CC-MAIN-2014-35 | 369 | 4 |
https://www.canada.ca/en/employment-social-development/programs/disability/savings/issuers/interface-standards/section5.html | code | Interface Transaction Standards: Technical Specifications
Disclaimer: RDSP issuers
The information contained on this page is technical in nature and is intended for Registered Disability Savings Plan (RDSP), Canada Disability Savings Grant (grant) and Canada Disability Savings Bond (bond) issuers. For general information, visit the RDSP section.
Consult this page frequently for newer versions. The following laws and regulations take precedence over information contained in InfoCapsules in the event of discrepancies:
- Income Tax Act
- Canada Disability Savings Act
- Canada Disability Savings Regulations
5.0 Technical Specifications
This section of the ITS document describes the data interface for the exchange of information between the CDSP system and issuers/authorized agents. These technical specifications are intended for use in support of system development to implement data interchanges with the CDSP system.
This document is the standard by which information is exchanged with the issuer/authorized agent for the application and administration of the grant and/or bond. Data integrity rules are described in detail in this document. Descriptions of the business and general rules under which data is processed within the CDSP system are described in the foreword of this document.
Operational aspects of the movement of data and functions used to manage the movement of data files are not part of this document but are found in the Data Operations and Connectivity Guide. Operational functionality includes the following:
- logging of files
- authentication of authorized agent
- transmission verification
- transmission mechanisms
Detailed operational instructions concerning reporting schedules and methods of transfer of information may be obtained by using the contact information below or by accessing the Service Partner page on the ESDC website.
- Electronic Services Section
- 140 Promenade du Portage, Mailstop: Bag 4
- Phase IV, Gatineau, Quebec
- K1A 0J9
- Telephone : 1-888-276-3632
- E-mail : [email protected]
5.2 CDSP System Transaction Processing Record Format Overview
These technical specifications describe a data interface that is based on the exchange of bulk data files. Authorized agents are required to conform to the record formats and rules specified herein as well as other data interchange rules described in the Data Operations and Connectivity Guide.
This portion of the ITS outlines both input and output data record formats. Input records are used to record contract registration information, record beneficiary and holder information and report financial transactions. Output records report the status of reported data in the form of transaction processing files and error files on a record-by-record basis.
5.3 Logical Record Types
Source transactions are identified by a record type code as outlined in the following table:
|001||Header Record (Source of transaction)|
|002||Sub-header Record (Used only when the CDSP system reports back to authorized agent in processing report)|
|003||Files Processed (Used only when the CDSP system reports back to authorized agent in processing report)|
|101||Contract Registration Information|
|102||Contract Update Transactions|
|201||Update Beneficiary and Holder or Add/Remove Holder|
|202||Add/Update and Revoke Consent Transactions|
|801||Transaction Errors in Error File|
|851||Severe errors in Error File|
|901||Successfully processed transaction in Transaction Processing File|
|921||SIN Usability information in SIN Usability File|
|951||Contract Status in Contract Status File|
|952||Episodic DTC Elections in Contract Status File|
|953||SDSP Elections in Contract Status File|
|971||Transferred information in Transfer Extract File|
|981||DTC Eligibility in Beneficiary DTC Eligibility File|
|999||Trailer Record (Control Count)|
5.4 File/Record Structure
- All transaction files have a header record containing standard identification details.
- All files have a trailer record containing a count of the number of records in the file including the header and trailer records.
- Files contain a mix of transactions, identified by a numeric record type code. This has been set at three (3) digits to allow for possible future expansion.
- Source input files contain fixed length records, with record types being padded as necessary to meet a consistent standard. This enables different record types to be included in the same file.
5.5 Data Formats
- The ISO-8859-1 Latin 1 Character Set is the official Treasury Board of Canada, Information Technology Standard (TBITS) for data interchange. All data is provided in ISO-8859-1 format (numeric values are stored in their character representation) as shown in Appendix B.
- All fields are fixed length and occupy fixed positions within a record.
- Character data is left justified and padded with trailing spaces except for BN.
NOTE: BN cannot be padded with spaces. If a record type "001" contains a space character (ASCII value 32) the file is rejected, and if any other record contains a space in the authorized agent BN field the record is rejected.
- Numeric data is right justified with leading zeroes.
- Most amount fields are standardized at 9 digits (10 bytes) with two explicit decimal places (i.e. up to a maximum of $9,999,999.99); negative amounts contain a minus sign "-" as the first character in the field.
NOTE: The amount field in record type "002" is longer than other amount fields allowing a maximum value of $9,999,999,999.99 (13 bytes).
- List-type data fields use code tables whenever practical (e.g. Province codes and Error codes).
- Record types 101, 102, 201, 202, 401, 501, 701 and 971 include a transaction type code. Separate codes are used with each type of transaction in order to identify the specific processing requirements.
5.6 Record Separators
Records within files must be separated by a record separator character(s). Record separator characters vary depending on the originator's operating system. The CDSP system replaces the carriage return (CR, decimal value 13) record separator character with the UNIX new line character (NL, decimal value 10).
Only the new line and carriage return characters are acceptable as record separators. No other record separator characters may be used.
5.7 End of File
The CDSP system rejects files that do not conform to the following rules:
- Files must have a type "999" record as the last record.
- The type "999" must have a record separator character following it.
If the end of file (EOF) character is provided, the following rules apply:
- The CDSP system accepts any single character as an EOF character following the type "999" record.
- No characters may follow the EOF character.
5.8 File Naming Standards
The physical naming of files is described as follows:
- To CDSP system : "CDSP " + File type + Authorized Agent BN + Transactions Latest Month + Date Sent + File number
- From CDSP system : "CDSP" + File type + Authorized Agent BN + Date Processed + CDSP system File number + extension
If the filename is not 36 characters long, and/or not formatted as outlined, an error record type 8001 is generated.
The combination of BN, Transactions Latest Month, Date Sent, and File number must be unique. If the same combination has already been received and processed by the CDSP system, the file is rejected and an error record type 8002 is generated.
If the Transactions Latest Month is in the future, the file is rejected and an error record type 8013 is generated.
The following definitions apply to the components of the file-naming standard:
|Program Identifier||The Program Identifier must be CDSP|
|File type||One uppercase character indicates the file type. The file type character indicates whether the file is a production file, a summary reporting file, a test file or a test summary reporting file.
|Authorized Agent BN||15 character Business Number|
|Transactions Latest Month||6 numeric character date YYYYMM|
|The latest month to which the transaction dates in the file relate|
|Date Sent||8 numeric character date YYYYMMDD|
|File number||2 digit file number|
|Must be between 01 and 99|
|Generated by CDSP system for .pro, .err, .sur, .reg and .xfr files.|
|Extension||The file extension is one of:
5.9 File Type
Production files being submitted to the CDSP system must begin with "CDSPP" while files starting with "CDSPT" are used strictly for industry testing and are never parts of a production file group. The procedures for industry testing are outlined in the CDSP System Industry Testing Guide.
5.9.1 File number
There may be instances where an authorized agent wishes to send more than one file in a single day. In order to be able to give each file a unique name, the file name contains a file number. If the authorized agent sends one file in a day, a file number must be provided, though it can be any two-digit value. The ordering of the file numbers will not be enforced. The file number is used purely to distinguish files sent on the same day.
5.9.2 File Extension
Files returned to the authorized agent have the same file type and BN but have the CDSP system processing date and file number in the prefix. In each reporting period a .pro , .err , .reg (if applicable), .xfr (if applicable) and .dtc file is returned to the authorized agent. The SIN Usability File, a .sur file, is returned to the authorized agent following the monthly production run. The following is an example of a filename group:
- Input Files
- Output Files
- CDSPP123456789RC00011998121501. err
- CDSPP123456789RC00011998121501. reg
- CDSPP123456789RC00011998121501. pro
- CDSPP123456789RC00011998121501. sur
- CDSPP123456789RC00012009121501. xfr
- CDSPP123456789RC00012009121501. dtc
All file names are in uppercase except file extensions.
5.9.3 Header and Trailer Record
The header record (adhering to the File Identification Standard) is the first record in the file and the trailer record, providing a control count of the records in the file, is the last.
The trailer record sent by the CDSP system contains:
- the unique file number assigned by the CDSP system and
- the date CDSP system processing occurred.
5.10 Source Data Definition Standard
Transaction format and content is defined in this document using a common ( COBOL ) standard, with the following symbols for data attributes:
|X||Any printable alphanumeric character (includes numbers, letters, punctuation marks, spaces and other special characters). The entire field contains spaces if not used. For example : the letter A in a 3 character alphanumeric field is stored as "A ". The number 5 in a character alphanumeric field is stored as "5 ".|
Any number. The entire field contains zeros if unused (blanks are not allowed). If larger than 1 digit, the contents are right justified with leading zeroes. For example : the number 5 in a 3 character numeric field is stored as "005".
Note: Negative amounts are preceded by a minus sign "–" as the first character in the field.
|( )||Indicates a recurrence of the preceding data type, with the number of occurrences stored inside of the parenthesis. For example : 9(6) means a number up to six digits long, X(6) means 6 consecutive characters of alphanumeric data.|
5.11 Standard Data Formats
The following table outlines standard formatting rules for common data field types:
|Dates||X(8)||Valid date formatted YYYYMMDD.|
|Up to maximum of $999,999,999.99. Decimals are explicit i.e. a contribution of $1000.00 is reported as 000001000.00 with the appropriate number of leading zeros for padding the field to the correct length (000001000.00).
|Filler||X(n-500)||Unused field. Must contain the specified number of spaces, or optional comments and is ignored regardless of its contents.|
All record types follow a standard layout, with the same fields occurring in the same positions to the extent possible.
5.12 Transaction Sequence
The CDSP system will process transactions in a logical order. Issuers should be aware of this and as much as possible all transactions should be submitted in a logical sequence.
All contract elements, including beneficiary, holder and PCG (where the beneficiary is under the age of 18) must be established in the CDSP system before financial transactions can be processed.
Contract registration information and the respective financial transactions pertaining to the contract can be sent in the same file. However, the processing of the financials remains dependent on the successful establishment of the contract elements.
Report a problem or mistake on this page
- Date modified: | s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875143373.18/warc/CC-MAIN-20200217205657-20200217235657-00157.warc.gz | CC-MAIN-2020-10 | 12,573 | 125 |
https://www.fhwa.dot.gov/cadiv/segb/views/document/sections/section4/4_8.cfm | code | - Briefing Room
U.S. Department of Transportation
Federal Highway Administration
1200 New Jersey Avenue, SE
Washington, DC 20590
This chapter describes typical systems engineering organizations and the role these organizations play in the development of ITS.
What makes an effective organization?
Effective systems engineering requires an integrated organizational structure with the following characteristics:
Because a system is being developed, the various disciplines that make up these teams, [For example hardware, software, or human-machine interface] are not independent of one another. Cross-coordination must be ongoing throughout project development. Continuing communication across disciplines is an essential function of the project organization for successful system development.
Specifically, the key criteria for an effective system management organization as adapted from Wilton P. Chase’s Management of Systems Engineering are:
The following is an explanation of each.
Few of the problems that arise in developing a system can be solved by a single discipline. Each provides a way of looking at the system. Complete understanding requires integrating these perspectives. This system view is an ongoing need. Therefore, the various team members must coordinate as the system is being developed. They must understand the viewpoint of the others and communicate in a language understandable to all.
A clear statement and understanding of the level of detail to be controlled at the project level makes management more efficient. It keeps the managers from slipping into too much detail emanating from their respective backgrounds. The process steps in Chapter 4.10 give guidance on how to tailor the process appropriately.
Simplify the paperwork
Standardized documentation is essential for efficient system management to record and transmit analyses, plans, and designs. During much of the systems engineering process, documentation is the only product. The system design is described only by specification. The following chapters of the Guidebook provide guidelines for developing documents appropriate to the scale and complexity of the project at hand.
Types of organizational structures
Functional One common approach is a functional configuration. Here each functional specialty or discipline is assigned to individual organizational entities. As an example, consider a systems engineering team who performs all systems engineering across all projects. This works best for small projects, where the team members may be working on several projects at once. Communications problems can occur for larger projects when sub-system teams are created. The risk is the sub-system teams may optimize for the sub-systems, not the system. Also, integration may be difficult since the pieces have been developed independently. This means that frequent cross-disciplinary communication and consideration of the system-level issues are essential.
Project The other approach is centered on projects, not disciplines. All those working on a project, no matter what their specialty, will report [possibly indirectly] to the project manager. This works only if the project is so large and long-term that the specialists can devote themselves to it for an extended period.
Matrix A hybrid approach, the matrix management structure, exists when team members report to both project and functional management. This approach is effective for large, long-term projects.
The Project Office approach, calls for project management, systems engineering, and design teams to be organized by project, and request project support from the functional staff as needed. This works for a moderate sized project, when only the key individuals devote full time to the project. The specialists work on multiple projects.
Integrated Product Team [IPT] this team consists of both agency and contractor representatives. They work together to develop the system that meets the project’s needs. In a large project, there are often mirror functions in the agency and contractor teams. For example, each has a program manager and a systems engineer. They work closely with their counterpart in the IPT. Further, representatives of each of the disciplines are a part of the IPT to ensure essential cross-discipline communication. Additionally, IPT’s may be formed to address key cross-discipline issues, such as cost of ownership, overall system performance, or configuration management.
Example organizational roles
Figure 4‑7 is an example of roles that are generally required for a successful systems engineering organization [adapted from Chase]. This may appear frighteningly complex, especially for an agency that typically does small projects. The important thing is that each box represents a role, not a department or an individual. A simple project may only require two people: a project manager and a systems engineer. Administrative functions assist on an as-needed basis. For larger organizations that manage more complex projects, this is a template for structuring groups with like activities together while maintaining system-level oversight and coordination.
There are three major activities in the organization: project management, systems engineering, and project control. Project management is concerned with planning and execution. Project control tracks the effort relative to performance, cost, and schedule goals. The same person may assume these two roles. Systems engineering is responsible for design, implementation, and verification.
Relationship to consultants and vendors
There is no single, correct, organizational structure. It needs to be tailored for each team based on existing structures and capabilities within the agency. It should take effective advantage of in-house expertise, existing working relationships, and communication paths. There are no standard roles for agencies and contractors. Agencies can [and often will] develop their own software, for example. Similarly, an agency may choose to outsource oversight activities. The only caveat is that there are certain activities which can only be performed by the agency. These key activities are listed for each step in the process throughout Chapter 3. The keys to a successful team that includes consultants are: appropriate roles and frequent, frank communications.
Figure ‑7 Example Organization | s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917119361.6/warc/CC-MAIN-20170423031159-00572-ip-10-145-167-34.ec2.internal.warc.gz | CC-MAIN-2017-17 | 6,380 | 27 |
http://hub05.hosting.pasteur.fr/vrouilly/ShinyLabExMICytometryBrowser/ | code | Upload my data to compare with LabExMI cohort.
In case you have acquired cytometry data following the LabExMI protocols, it is possible to upload here your data so that they can be projected on top of the results collected on the 1000 healthy donor cohort of the LabExMI.
example CSV file
- Export your cytometry statistics data as a CSV file (with FlowJo for example):
- make sure it uses a comma as a separator (”,”)
- the first line of the file is expected to be the columns’ header. It needs to at least list the phenotypes of interest. Only phenotypic names matching exactly the LabExMI phenotypes will be considered.
- add 5 new columns to your data with group information: AgeGroup (20-29yr,60-69yr), GenderGroup (Men, Women), AgeGenderGroup (20-29yr M, 20-29yr F, 60-69yr M, 60-69yr F), AGE.V0 (Age in years i.e 22.7), CMV.V0 (Positive/Negative)
- one row should represent one donor/patient
- other columns are permitted in the file but will be ignored.
- Use the upload button to select the file you wish to analyse.
- Following the upload, the “boxplot view” will be set automatically so that the uploaded phenotypes will be presented all at once, and the uploaded data will appear on top of the LabExMI data, as black dots.
- An example CSV file is available at the bottom of this page.
- you can contact me at [email protected] if you have any questions, or comments. | s3://commoncrawl/crawl-data/CC-MAIN-2023-50/segments/1700679100674.56/warc/CC-MAIN-20231207121942-20231207151942-00403.warc.gz | CC-MAIN-2023-50 | 1,402 | 13 |
https://designshack.net/design/lance-charlton/ | code | This is the portfolio of Lance Charlton, a UI / Graphic Designer from Southern California. For 9 years he has been designing and developing user interfaces and web applications for companies large and small. This site relies heavily on CSS3 and jQuery.
About Lance Charlton
A selection of other designs you might like to explore, based on Lance Charlton: | s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370505359.23/warc/CC-MAIN-20200401003422-20200401033422-00361.warc.gz | CC-MAIN-2020-16 | 354 | 3 |
https://boards.weddingbee.com/topic/need-some-help-to-pick-a-new-laptop/ | code | - 6 years ago
- Wedding: October 2014
SOO….my laptop bit the big one, mother board fried! : (! so now i need a new laptop, for continuing my education and general use, the only thing my program says (online classes) is that they reccomend and intel processor (i have no clue if it is better than adm or what is good now days!), so we are looking at laptops, i havent had to buy a laptop in about 4 years, so it is time to upgrade my little acer (before it fried), soo ladies…what kind of laptops do you have?? i will need it to do online class work, wedding planning, and general computing! Really appreciate any advice you guys can give me….we are hoping to see if we can stay under the $500 mark…. | s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267156314.26/warc/CC-MAIN-20180919235858-20180920015858-00341.warc.gz | CC-MAIN-2018-39 | 707 | 3 |
https://leapfrog.nl/blog/archives/2008/06/06/moving-speaking/ | code | It’s final days for me. In Copenhagen, that is. July 1 I will exchange this lovely city for my home town of Utrecht, the Netherlands. The plan is to continue work as a freelance interaction designer. So if you’re interested, but physical distance has been putting you off so far, get in touch.
Between now and then, most of my time will be spent at conferences. Here’s the rundown:
- First up is From Business to Buttons, June 12–13 in Malmö, Sweden. My talk is titled More Than Useful. I will attempt to show that for a certain class of products, playfulness is a vital characteristic. The idea is to introduce the IxD crowd to some game design concepts.
- The week after that I will be at the Festival of Games, June 18–20 in Utrecht, Netherlands. My presentation is titled Playing With Complexity. I will introduce the game design audience to some interaction design thinking and suggest data visualization might be an interesting area to team up on.
- Last but not least is good old Reboot, 26–27 June in Copenhagen. I have submitted a proposal titled Playful Activism in the Real-Time City, which I hope will be selected to be on the program.1
If you will be at any of these conferences, do drop me a line or say hello at the event itself.
- If you’d like to see it too, don’t hesitate to vote it up. [↩] | s3://commoncrawl/crawl-data/CC-MAIN-2024-18/segments/1712296817184.35/warc/CC-MAIN-20240417235906-20240418025906-00705.warc.gz | CC-MAIN-2024-18 | 1,328 | 7 |
https://www.experts-exchange.com/questions/28375988/Software-or-solution-for-file-searching-organizing-to-work-with-LAN-shared-folders-as-well.html | code | I have a company business partner who recently bought few file servers with 8 and 16 HDDs x 4 TB. The primary business goal is to keep on them video files – so they need fast search and access the servers /video files/ primary via LAN. What software tool you will recommended me – which organize the video files in catalog, so when they need to search a file by its name to avoid the need to manually click on all shared directories?
Both PC - the file server and the PCs used Win 7.
A free solution will be prefered, anyway shareware solution may be accepted as well. | s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514573052.26/warc/CC-MAIN-20190917040727-20190917062727-00091.warc.gz | CC-MAIN-2019-39 | 572 | 3 |
https://triplebyte.com/company/public/dirt-protocol/VaqJMiA/back-end-engineer | code | - Silicon Valley, CA, United States
About DIRT Protocol
DIRT is a protocol for trusted data. Our mission is to build the Wikipedia for structured data, and make it freely accessible. Today, we depend on a few gatekeepers for reliable data. This system falls apart if the sources are reliable or have an incentive to lie. Rather that trusting the source for the information, DIRT defines a set of rules that incentivize honesty and make it economically irrational for misinformation to persist.
Founded this year by Stanford and Y Combinator alumni who sold her last company to MSFT. DIRT raised $3M from Greylock, General Catalyst, SV Angel, Lightspeed, Pantera, etc. Join a small, agile, and experienced team with three YC alumni to build the data layer for the blockchain.
We are an experienced team with three Y Combinator alumni. We've started and scaled companies before. We operate with a high degree of autonomy and prioritize shipping products to maximize learnings.
As a member of the team:
You'll get autonomy to make decisions AND mistakes. We're a team that biases towards shipping code and learning.
You'll find a team that trusts each other. We step up for each other to create space for the team to do their job.
You'll learn how to express your thoughts. We actively work on clear and succinct communication of ideas
You'll learn to build decentralized apps and be a early thought leader in the space.
Skip straight to final-round interviews by applying through Triplebyte. | s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655906214.53/warc/CC-MAIN-20200710050953-20200710080953-00050.warc.gz | CC-MAIN-2020-29 | 1,489 | 11 |
https://cofounderslab.com/discuss/can-a-python-elastic-stack-handle-big-data-2-billion-records | code | Lead data diagnostics at Vibe
We are estimating our product to be aggregating around 2 billion data points in the next year. Each data point will be a combination of 15 to 20 information items (text) about entities. We use Python/Django for the application and APIs and Elasticsearch as the data store.
Will this stack scale to 2 billion records and processing? If no, what should I be considering. Would appreciate your suggestions. | s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964363520.30/warc/CC-MAIN-20211208175210-20211208205210-00227.warc.gz | CC-MAIN-2021-49 | 433 | 3 |
https://practice.geeksforgeeks.org/courses/sql?utm_source=article&utm_medium=article&utm_campaign=courses%2Fsql&utm_id=https%3A%2F%2Fpractice.geeksforgeeks.org%2Fcourses%2Fsql | code | Structured Query Language or SQL
is the programming language for relational databases like MySQL, Oracle, Sybase, SQL Server, Postgre,
and more. As the name suggests, it is used when we have structured data (in the form of tables). So if your objective is to store, retrieve and maintain data in a structured manner, SQL comes in very handy.
In todays times of internet, there is no business that doesn't use or store data in different databases, and so in order to communicate with these databases, you need to learn SQL. It is an integral part of business operations,
that is why learning this skill will always be beneficial for you. Also, in order to advance your career in the field of Data Science
, SQL is one of the basics that you must learn Learn more about how learning SQL can help your career grow
This SQL Self Paced course will teach you the basics of databases, types of commands, DML commands, functions
and all other concepts of SQL that are a must know in the field of data science. Along with the great explanations of the concepts, you also get access to assessment tests
that will help you judge your progress. | s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233510238.65/warc/CC-MAIN-20230927003313-20230927033313-00721.warc.gz | CC-MAIN-2023-40 | 1,132 | 9 |
https://cs-people.bu.edu/yzxie/cs585/homework3.html | code | CS 585 Homework 3
Yize Xie U14485891
- Implement a connected component labeling algorithm.
- Apply a technique to reduce the number of components.
- Implement the boundary following algorithm.
- For each relevant region/object, compute the area, orientation, circularity (Emin/Emax), and compactness.
- Implement a skeleton finding algorithm.
- Implement two segmentation algorithms.
- Apply the tools you implemented for part 1 to analyze your data.
Method and Implementation
Connected Component Labeling
We use Deep First Search(DFS) algorithm for connected component labeling, and we implement it by using stack.
- ensure image is binarized.
- if stack is empty goto 3, if not empty, goto 4.
- find an unprocessed pixel and push it to stack, and mark it as a new component. If all of pixels have been processed, terminate.
- pop the front element of stack, push neighbors of it into stack (if not processed), and mark them the same as the front element.
- go back to 2.
If the component we find has too low area, more specifically, lower than a threshold we set, the component will be discarded.
We implement Moore-Neighbor Tracing with Jacob's stopping criterion, which is very good for boundary detection on 8-connectivity components. In the algorithm, we keep track with an inner pixel and an outer pixel. The pointers move clockwise until they are back to beginning in a same manner.
Our skeleton finding algorithm belongs to thinning. We erode the image gradually with a cross kernel, and what remains is exactly the skeleton we need to find.
Experiments and Results
Below is sample of Fist and Palm.
As is shown in the above images, our algorithm detects component, boundary and skeleton successfully. For orientation, it is degree anti-clockwise from the axis "top to bottom". Reasonably, fist has a relatively higher circularity than palm.
White balance: We want to conduct skin detection later, but the orange light in the hall will make everything look like skin. Therefore, firstly we need to white balance images so that everything will look naturally.
Compute background: Identifying what is moving will clean the image to a great extend. Therefore, we compute the background by combining part of two frames. This should be acceptable in real world scenario because we can take a photo before musician comes.
Get difference: For each frame, calculate the difference from background and here we conduct segmentation using absolute threshold algorithm.
Recover color: Thresholded image is gray scaled, but what we need to do skin detection should be colored. We use the binarized image as a mask on original one to get color.
Skin detection: Skin detection is based on HSV value, and in order to deal with shadow, we detect an extra one with brightness augmented. Morphology OPEN and CLOSE are used here to reduce noise. The circularity of hands is typically from 0.2 to 0.7.
Identify two hands: The last step is to classify which one is top hand and which one is bottom. The trick here is that two hands should take similar space in expectation.
- Compute background: On the gray scale image, there are many noise in the bottom part. But pattern here is that bats are always brighter than background. So we simply calculate min value of each frame as the background of this dataset. Then we
- Get difference: We use absolute difference to extract bats from original images.
- Absolute threshold: Absolute threshold here is enough to extract bats from small noise and have a good result.
- Adaptive threshold: Absolute threshold is used to detect heads of bats (extracting them from wings). It is useful because later we need to decide whether there are multiple bats in a single component.
- CCL: After reducing noise with OPEN and DILATION, we conduct connected component labeling. Each component represents a bat.
- Process exception: The exception occurs when multiple bats overlap. Our assumption here is that in the heat map, the head of each bat has greater brightness than other part. So if we detect multiple heads in the same component, we will draw thicker rectangles to represent multiple bats.
- Calculate circularity: if the circularity is higher, the rectangle will be whiter and the bat is more likely to hold its wings. On the contrary, if the circularity is lower, the bat is more likely to expand its hands.
- Compute background: In this scenario, we expect the background image to be the averaged frame. However, due to two people chatting in the middle for long time, the average value is not accurate here, so we use some part of first frame to supplement middle part. It should be acceptable in real time because if we have more data, the averaged frame should be more accurate.
- Get difference: We use absolute difference to extract pedestrians from original images. However, there are too much noise here.
- Segmentation: We use adaptive threshold here to extract humans. Although we already selected best parameter, morphology like OPEN and CLOSE are still conducted to reduce noise.
- CCL: This image, although noise still in it, is best we can get before connected component labeling. We conduct CCL and afterwards, filter components that are too small to be a human. Regions with width > 15 and height > 37 are typically kept by us, because they look like human.
- Object tracking: We have simple but workable object tracking algorithm here (without any OpenCV object tracking function). That is we assume the component and the component detected in the previous frame which are close are from same object. We mark same object with same color.
We get very good result on every dataset. But if we have more time, we should spend some on trying to detect human behind obstacles, or extract humans from each other when they are close.
We are very happy to find that our bat detection algorithm can detect some bats that is really hard for human to observe.
We learn really a lot from this homework.
Credits and Bibliography
My teammates are Wenxing Liu and Weifan Chen. | s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610704821253.82/warc/CC-MAIN-20210127055122-20210127085122-00705.warc.gz | CC-MAIN-2021-04 | 6,010 | 46 |
https://egghead.io/lessons/git-update-my-feature-branch-with-the-latest-changes-from-master | code | Instructor: Our situation in our repository looks as follows. I'm currently working on that app refactoring feature branch, where I've added two commit messages. Then on master, there are other commits on top of it, which are not yet in my feature branch.
In this case, I have created them on purpose, but it might be that your teammate has created them, pushed them to the remote repository. You did a git pull today morning, and you got down those commit messages in your master branch.
As in this example here, it's a good practice to create a feature branch for new functionality you create, or even back fixes. However, I highly recommend you to synchronize it with the main development line, which might be master or some dedicated branch, as often as possible.
The main reason is that you can adapt immediately to changes which other developers made to the main application code. Also, in this way, you avoid really, really large merge problems in the end when your branch is finished.
To update our code, first of all, we need to pull down everything on master. We need to make sure that we are synchronized and up-to-date. Then we just jump to our app refactoring branch. Now, in order to synchronize with master, we have different kind of possibilities.
You can obviously do a simple git merge master, which would then create a new commit, which is the result of merging the changes which happened in master into our app refactoring branch. This is highly recommended, especially if you have a share branch. If you pushed that already to the origin, and other developers are working that as well.
However, if this is your reserved feature branch, where just yourself is working on, then you can use a command which is git rebase. What git rebase will do is to take each commit from master and merge it on top of your commits in your feature branch.
As a result, our feature branch will simply be shifted upwards, as if we just created it out of the latest version of the master branch. However, as you can imagine, that changes these commit hashes. Again, only do that if it's your own personal branch, and it is not shared with others.
Let's perform the rebase. We do a git rebase with master. As you can see, what it does, it is rewinding the hat, and replace our work on top of master. If I do now a git log, we can see that we have now a linear line in our Git commit history.
It looks like we just created this app refactoring branch and added two of these commits. However, these now also contain all of the latest changes that have happened in master. | s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370521876.48/warc/CC-MAIN-20200404103932-20200404133932-00016.warc.gz | CC-MAIN-2020-16 | 2,568 | 10 |
http://metabunk.org/threads/326-Debunked-Human-Protection-Software-Suite?p=9535&viewfull=1 | code | I've titled this "debunked" to aid with searching, but really this is something that should not need debunking, as it's obvious nonsense:
So the question here is: is this actually serious? It's marketed at the "TI" (Targeted Individual) community (Gang Stalking), which certainly exists, as a community. But to regular folk this seems like a parody. Is this just someone having cruel fun with the clinically paranoid TI community? Or is the vendor, Mell Mellhedek, really that deep down the rabbit hole?We have a full solution to reach a full Human Protection. Our solution is automated but it doesnīt cover 100% of the device removal. That means that there are devices(also known as implants or chips) that our solution will not find or will not remove automatically because of its size or its implications. At this moment those devices that were not removed automatically we can remove them manually (human assisted), which means man/hour fees. There are also cases that some device might not be removed because of the damaged done to the human body. Sometimes a physical part has been substituted by a device, so the removal would result in a null part in the body.
Our solution has already been successfully tested in Brazil and USA.
Our Human Protection Software Suite is a composition of many softwares installed in the Humanīs body. After its installation, it starts working with the following components:
- Human Protection Dome - impedes new device implants (installation) from entering your body
Our solution is under constant development and will be kept that way to keep ahead of the otherīs tech.
- Quena Human Shield - reflects outside emissions.
- Human Firewall - impedes the backdoor connections to your body for influencing, dream injections and information leakage.
- Automatic Device Removal Matrix - removes devices (chips, implants) from the body in a search and remove algorithm. The removal is done in the first 7 layers of the human body, and recovers the damaged area using Human Technology for Recovery. It removes from milli, micro, nano scale down to sub-nano scale devices. The bigger devices are imploded but not all will disappear, thou will stop working.
- Program Tracker - searches and removes programs and configurations in a search and remove algorithm.
- Anti Stalking Tech - creates a 2Km (aprox. 1.2miles) clearance area around you removing stalking intentions against you.
- Embedded Human Technology - used to recover the damaged areas from the device removal
- Anti Tracking - hides a human in the environment, so that tracking possibility is null to their tech.
We do track you via your bio signature and test you for compatibility with our tech. So not everyone will be accepted, and we will only accept those that meet compatibility with our tech.
Our Human Protection Software Suite will be licensed by a months/fee as soon as we reach open market. | s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368697974692/warc/CC-MAIN-20130516095254-00008-ip-10-60-113-184.ec2.internal.warc.gz | CC-MAIN-2013-20 | 2,898 | 15 |
https://fm-alchemist.livejournal.com/3962100.html | code | Random Drabble- Takes place at the end of series, before the movie.
Word Count- 357
I don’t own Fma.
An angel had spoken to me once.
Yea, you think I’m kidding? I thought it was a joke myself! I don’t believe in religious crap, but I swear an angel one day flew from the sky and sat by my window. (I had to look twice...) | s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107867463.6/warc/CC-MAIN-20201019232613-20201020022613-00093.warc.gz | CC-MAIN-2020-45 | 327 | 5 |
https://www.karkidi.com/Find-Jobs/Category/Senior%20Data%20Visualization%20Manager | code | 19 Aug 2022
We are the 100% digital bank of the Santander Group and we are currently undergoing a technological transformation and international expansion. In 2016 the re-launch of the Bank began and since...
Java Programming,Design,Management,Python Programming,SQL,AWS,Big Data Technology,Scala Programming,Machine learning techniques,Jenkins,JIRA,Data Visualization,NoSQL,Cloudera,Amazon Athena,Postgres,MySQL,Amazon RedShift,Amazon EMR,SPARK Programming,Database testing,Lambda,Amazon Simple Storage Service (S3),Apache Flink,Splunk,Data Warehousing
This is a unique opportunity to join the Financial Crime Data Science and Analytics area to lead the Strategic Transaction Monitoring team of Data Scientists and Engineers. Our purpose is to...
Boadilla del Monte, Spain
Santander Consumer Mobility Services (Ulity), the new mobility company of Santander Group.
Would you like to be part of one of the largest transformations that the World is...
Java Programming,Risk,Python Programming,SQL,PowerBI,Big Data Technology,Apache Hadoop,Scala Programming,Machine learning techniques,Data science techniques,Data Visualization,Tableau,Algorithms,SPARK Programming,Bash scripting
Boston, MA, USA; New York, NY, USA; Pennsylvania, USA; New Jersey, USA; Miami, FL, USA
Java Programming,Analytical and Problem solving,Leadership,Python Programming,SQL,Machine learning techniques,Data science techniques,JIRA,Data Visualization,Data Modeling,Tableau,R Programming,Statistics,SAS,Data Engineering
Dallas, TX, USA
The Data Science Analyst (Sr. Specialist, Pricing) is responsible for analysis, sales and pricing of a large portfolio of products. Key activities are to develop consistent and reliable reporting...
Wordpress,Java Programming,Risk,Python Programming,SQL,Data science techniques,Powerpoint,Data Analysis
Santander, Spain; Glasgow, UK
As Product Owner within Santander’s Data Science area, you’ll own, articulate, and oversee the delivery of a clear product vision and roadmap to increase our Machine Learning...
Java Programming,Analytical and Problem solving,Risk,Python Programming,SQL,Machine learning techniques,Data science techniques,Data Modelling | s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882573667.83/warc/CC-MAIN-20220819100644-20220819130644-00608.warc.gz | CC-MAIN-2022-33 | 2,164 | 16 |
https://www.torrentdownloads.me/torrent/879841/riplinux+2+2 | code | Recovery Is Possible (RIP) is a Slackware-based CD or floppy boot/rescue/backup/maintenance system. It has support for a lot of filesystem types (Reiserfs, Reiser4, ext2/3, iso9660, UDF, XFS, JFS, UFS, HPFS, HFS, MINIX, MS DOS, NTFS, and VFAT) and contains a bunch of utilities for system recovery. It also has IDE/SCSI/SATA, PCMCIA, RAID, LVM2, and Ethernet/DSL/cable/PPP/PPPOE network support.
Updated packages: cdrskin v0.3.4.pl00, parted v1.8.3, v1.5.14+cvs20070315,
xchat v2.8.0, xarchiver v0.4.6, msmtp v1.4.10, rsync v2.6.9, tin v1.9.2,
fluxbox v1.0rc2 and a few other updates!
This X version will be the only version of the Linux system. | s3://commoncrawl/crawl-data/CC-MAIN-2016-50/segments/1480698543567.64/warc/CC-MAIN-20161202170903-00019-ip-10-31-129-80.ec2.internal.warc.gz | CC-MAIN-2016-50 | 645 | 5 |
http://battleofbrothers.com/sirryan/piecing-it-together/ | code | This post is a status update, which is how I track progress specific to my game. Check out my timeline to see what I've recently accomplished.
For the past year I’ve been working on a bunch of different code bases learning new concepts. For this update, I decided to piece together a first build of my “game” to get rid of redundant bits of code. I’m left with an app that has menus, navigation, saving, loading, and cutscenes. While it’s still far from an enjoyable game, it’s cool stuff to have a foundation to work with.
To see everything in motion, check out the video below.
Bringing Cutscenes To The Next Level
The original plan for this update was to do another round of polish on the story. Fortunately, I had a friend in town — Brian from Secret Powers. He specializes in trailer / movie making / special effects, so was able to provide just the critique I needed. He actually came up with the gif below to show how much better the story could be.
Now that I’ve seen the true potential, I have a better idea of what direction to go. Also, I’m considering working with Brian on trailers and cutscenes for the game because the improvement is so noticeable.
Foundation of a Game
Because I’m on hold until I find out to what extent I’ll be working with Brian, I decided to move back to code. I took each of my separate projects (story, battle, tests) and started merging them into one code base. That provided the appearance of a game as you can see below:
The cool part about this is that the following now work reliably:
- Scene management and caching
- Conditional scene ordering
- Data accessors (i.e: what troops are currently in my army)
- Basic UI framework (propagating buttons, action callbacks, etc)
I haven’t finished bringing the battle into this project yet, so that is where I’ll pick up. Once that is complete, it will feel like I have a real project that I’m building off of.
Other Things in Motion
I’m moving on to my last week with Scott, our illustrator. With this completed, design and thought has been put into all maps, merchants, heroes, upgrades, camps, and units. I’m in a fairly good spot, and even though it is over budget the time restriction on this contract has been nice to enforce scope.
Aside from that, I’ll be starting music in a couple of weeks, and I have postponed voice because of the story uncertainty mentioned above. Nothing is planned for sound yet. My rough thinking is that when music is done, I’ll just focus and get a fun, functional game working. Then, I’ll revisit working with contractors to add that much needed polish. | s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233510603.89/warc/CC-MAIN-20230930050118-20230930080118-00589.warc.gz | CC-MAIN-2023-40 | 2,613 | 17 |
https://eden.sahanafoundation.org/wiki/BluePrint/DeploymentTools | code | BluePrint: Deployment Tools
Table of Contents
The project aims to develop an Eden application that allows deployment of Sahana to production servers. These may be Amazon EC2 instances or generic Linux Servers. The application would provide flexibility during deployment. For instance, the deployer may be given an option to select the Web Server, Database, Mail Server and Co-Apps of their choice. Further support for Amazon EC2 can be incorporated which would allow the creation of new instances and Eden deployment on the newly created instances. The application can further evolve to cover WebSetup for deployed instances.
- Users who prefer a graphical interface over the command line, particularly those who have less experience with Linux.
- System Administrators, Site Operators
A deployer with minimal experience with linux and command line wants the application to deploy Sahana Eden (from trunk or personal github repo) so that they may use it in production
The application is expected to:
- Be flexible with deployment options – the end user should have the choice to select the web server, database and mail server of their choice. Further, a set of advanced options may be added which allow users to specify the github branch to clone repo from, specify the template after deployment, etc.
- Integrate well with Amazon EC2. The application shall allow the user to create new free/paid amazon instances using the amazon cli tools in the backend while allowing them to deploy on already existing instances as well.
The application can use Fabric (http://fabric.org) in the backend to do the work behind the scenes.
Incorporating the WebSetup as an Eden app as a part of this project, though ambitious, would make it easier to maintain and integrate. | s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882571847.45/warc/CC-MAIN-20220812230927-20220813020927-00049.warc.gz | CC-MAIN-2022-33 | 1,762 | 11 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.