id
stringlengths 32
32
| text
stringlengths 0
895k
| name
stringlengths 0
33k
| domain
stringlengths 5
44
| bucket
stringclasses 19
values | answers
list |
---|---|---|---|---|---|
87285b95e1f72d7a0366b604f6e32b5e |
Below are ours server , xlog & archive details
```
PRIMARY / MASTER:
================
postgres@tutorialdba\_1:> ps -ef|grep postgres
postgres 4436 1 0 Jun25 ? 00:05:04 /opt/10.6/bin/postgres -D /data
postgres 4437 4436 0 Jun25 ? 00:00:00 postgres: logger process
postgres 4439 4436 0 Jun25 ? 00:10:19 postgres: checkpointer process
postgres 4440 4436 0 Jun25 ? 00:00:40 postgres: writer process
postgres 4441 4436 0 Jun25 ? 00:04:55 postgres: wal writer process
postgres 4442 4436 0 Jun25 ? 00:01:31 postgres: autovacuum launcher process
postgres 4443 4436 0 Jun25 ? 00:00:27 postgres: archiver process last was 00000001000000450000005E
postgres 4444 4436 0 Jun25 ? 00:06:55 postgres: stats collector process
postgres 4445 4436 0 Jun25 ? 00:00:02 postgres: bgworker: logical replication launcher
postgres 4500 4436 0 Jun25 ? 00:05:07 postgres: wal sender process repmgr 53.99.98.119(44112) streaming 45/5FF3DB10
XLOGS :
postgres@tutorialdba\_1:/xlog>
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 000000010000004500000096
-rw------- 1 postgres postgres 16777216 Jul 22 04:01 000000010000004500000097
-rw------- 1 postgres postgres 16777216 Jul 22 04:01 000000010000004500000098
-rw------- 1 postgres postgres 16777216 Jul 22 04:01 000000010000004500000099
-rw------- 1 postgres postgres 16777216 Jul 22 04:01 00000001000000450000009A
-rw------- 1 postgres postgres 16777216 Jul 22 04:01 00000001000000450000009B
drwx------ 2 postgres postgres 36864 Jul 22 05:01 archive\_status
XLOG STATUS :
postgres@tutorialdba\_1:/xlog/archive\_status> ls -l
total 0
ARCHIVE LOGS:
postgres@tutorialdba\_1:/archive\_log>
-rw------- 1 postgres postgres 16777216 Jul 22 04:01 000000010000004500000059
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005A
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005B
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005C
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005D
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005E
drwx------ 2 postgres postgres 20480 Jun 14 13:55 archive\_status
STANDBY / SLAVE :
=================
postgres@tutorialdba\_2:/archive\_log> ps -ef|grep postgres
postgres 4488 1 0 Jun25 ? 00:00:57 /opt/10.6/bin/postgres -D /data
postgres 4489 4488 0 Jun25 ? 00:00:00 postgres: logger process
postgres 4490 4488 0 Jun25 ? 00:16:08 postgres: startup process recovering 00000001000000450000005F
postgres 4493 4488 0 Jun25 ? 00:10:07 postgres: checkpointer process
postgres 4494 4488 0 Jun25 ? 00:00:33 postgres: writer process
postgres 4495 4488 0 Jun25 ? 00:01:08 postgres: stats collector process
postgres 4496 4488 0 Jun25 ? 00:44:00 postgres: wal receiver process streaming 45/5FF3DBF0
XLOGS :
postgres@tutorialdba\_2:/xlog>
-rw------- 1 postgres postgres 16777216 Jul 22 04:01 000000010000004500000059
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005A
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005B
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005C
-rw------- 1 postgres postgres 16777216 Jul 22 04:02 00000001000000450000005D
-rw------- 1 postgres postgres 16777216 Jul 22 20:39 00000001000000450000005E
-rw------- 1 postgres postgres 16777216 Jul 22 12:36 00000001000000450000006F
drwx------ 2 postgres postgres 970752 Jul 22 04:02 archive\_status
XLOG STATUS :
postgres@tutorialdba\_2:/xlog/archive\_status>
-rw------- 1 postgres postgres 0 Jul 22 04:01 000000010000004500000057.done
-rw------- 1 postgres postgres 0 Jul 22 04:01 000000010000004500000058.done
-rw------- 1 postgres postgres 0 Jul 22 04:01 000000010000004500000059.done
-rw------- 1 postgres postgres 0 Jul 22 04:02 00000001000000450000005A.done
-rw------- 1 postgres postgres 0 Jul 22 04:02 00000001000000450000005B.done
-rw------- 1 postgres postgres 0 Jul 22 04:02 00000001000000450000005C.done
-rw------- 1 postgres postgres 0 Jul 22 04:02 00000001000000450000005D.done
-rw------- 1 postgres postgres 0 Jul 22 04:02 00000001000000450000005E.done
Archive log location :
postgres@tutorialdba\_2:/archive\_log> ls -l
total 0
```
| How to DELETE current XLOG / WAL LOG in postgresql database ? | tutorialdba.com | 2019.43 | [
{
"text": "\n**Solution :**\n\n\n \n\nSlave win't be archive if even archive enabled as well.\n\n\n \n\nCheck the status of xlog then delete the .done files before that check the master xlog status, walsender and last archived log for your confirmation.\n\n\nso finally you can delete the below xlog untill (000000010000004500000059 - 00000001000000450000005E). \n\n \n\n\n",
"name": "",
"is_accepted": false
}
] |
840bd279d991ae8416158075a80903f4 | How will run any shell script in background? | tutorialdba.com | 2021.43 | [
{
"text": "\n By giving\" &\" at end of script called running background\n\n\n (OR)\n\n\nSet in task scheduler or [crontjob](https://www.tutorialdba.com/2017/11/what-is-crontab-and-how-to-work-with.html?m=1).\n\n\nUsing at command you can schedule a job to run at a particular date and time. For example, to execute the backup script at 5 a.m tomorrow, do the following.\n\n\n\n```\n$ at -f backup.sh 5 am tomorrow\n```\n\n**you can run the shell script Following methods at backgroundly:**\n\n\n\n```\n. script.sh &\n./script.sh &\nsh script.sh &\n (OR)\nnohub sh script.sh &\nnohub ./script.sh &\nnohub .script.sh &\n```\n\n**Note: hup** is a hangup signal... nohup - no hangup... so if we give nohup script.sh & ... it starts progressing script.sh in back ground... that the subsequent logout or session disconnection does not stop..\n\n\n**Note:**& Stands for running background\n\n\nAfter running this give \"ctrl+z\" in background one id will be displayed as running that is called process\\_ID using that PID we can adjust [priority](https://www.tutorialdba.com/search/label/Linux%20Priority%20Changing?m=1) for that particular jobs.\n\n\njobs -l (will display that process id which is running )\n\n\n**Note**: jobs will be displayed only the running and stopped jobs for only that session\n\n\n[top](https://www.tutorialdba.com/2017/08/linux-top-command.html?m=1) -c is also one of option -c means which display the commands. But This will be displayed whole server jobs and services not for session only jobs.\n\n\nif you want commands\n\n\n\n```\nscreen \nSh script.sh\n```\n\nHere are the steps you can follow to run a process in screen, detach from the terminal, and then reattach.\n\n\n1. From the command prompt, just run ***screen***. This will give you a new subshell.\n2. Run your desired program\n3. Detatch from the screen session using the key sequence **Ctrl-a Ctrl-d** (note that all screen key bindings start with Ctrl-a). This will drop you back to your original shell and display a message \"[detached]\", indicating that the screen session is still running.\n4. You can then list the available screen sessions by running \"**screen -list**\"\n5. You can reattach to this screen session by running \"**screen -r**\". Once reattached, you will be able to take off where you left off and see any output that was printed to the screen during the time that you were detached. If you have multiple screen sessions, then you can specify the tty name (as displayed by screen -list) as an argument to screen -r to attach to a particular session.\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
4062f66ec1a1801cb45c6979877c5a6b | The Most Beautiful Royal Women in the World: queens and princesses ? | tutorialdba.com | 2020.34 | [
{
"text": "\nIn the modern world is developing and improving, the monarchy continued to exist, sometimes having a huge impact on the development of their countries. Many royal women are widely popular among his people, and are also involved in social, political and charitable activities. Many of the modern princesses and queens had never belonged to the royal couple, and married a member of the monarchist family, thus receiving the royal title. In the Top-16 Most Beautiful Royal Women in the World: queens and princesses came from 13 countries: Bhutan, Nepal, Morocco, Norway, Luxembourg, Sweden, Spain, the UK, India, Egypt, Saudi Arabia, Jordan and Monaco. 16. Queen Jetsun Pema (born 4 June 1990) is the Queen consort (literally meaning \"Dragon Queen\") of Bhutan, as the wife of King Jigme Khesar Namgyel Wangchuck. 15. Mette-Marit, Crown Princess of Norway (born 19 August 1973 in Kristiansand, Norway), is the wife of Crown Prince Haakon, heir apparent to the throne of Norway. See also: The most beautiful Norwegian women 14. Himani Shah, Former Crown Princess of Nepal (born October 1, 1976) is the wife of the former heir apparent to the throne of Nepal, Crown Prince Paras. 13. Princess Lalla Salma of Morocco (born Salma Bennani; 10 May 1978 in Fes) is the princess consort of Morocco. She is the wife of King Mohammed VI and the first wife of a Moroccan ruler to have been publicly acknowledged and given a royal title. See also: The most beautiful Oriental women 12. Princess Alexandra of Luxembourg (born 16 February 1991) is the fourth child and only daughter of Grand Duke Henri and Grand Duchess Maria Teresa. 11. Princess Madeleine of Sweden (born 10 June 1982), is the youngest child and second daughter of King Carl XVI Gustaf and Queen Silvia of Sweden. Upon her birth, she was third in line of succession to the Swedish throne. After the birth of her niece, Princess Estelle, in February 2012, she became fourth in line of succession. See also: The most beautiful Sweden women 10. Catherine, Duchess of Cambridge (\"Kate\" Middleton; born 9 January 1982) is the wife of Prince William, Duke of Cambridge. Following his father Charles, Prince of Wales, William is second in line to succeed his grandmother, Queen Elizabeth II, as monarch of 16 Commonwealth realms. 9. Letizia, Princess of Asturias (born 15 September 1972), is the wife of Felipe, Prince of Asturias, the heir apparent to the throne of Spain. Before her marriage to the prince, Letizia Ortiz Rocasolano was a journalist and news anchor. As consort to the heir apparent to the Spanish throne she bears the official titles of Princess of Asturias, Princess of Gerona, Princess of Viana, Duchess of Montblanc, Countess of Cervera and Lady of Balaguer. See also: The most beautiful Spanish women 8. Diana, Princess of Wales (1 July 1961 – 31 August 1997), was the first wife of Charles, Prince of Wales, who is the eldest child and heir apparent of Queen Elizabeth II. 7. Gayatri Devi (23 May 1919 − 29 July 2009), often styled as Maharani Gayatri Devi, Rajmata of Jaipur, was born as Princess Gayatri Devi of Cooch Behar. She was the third Maharani of Jaipur from 1939 to 1970 through her marriage to HH Maharaja Sawai Man Singh II. 6. Princess Haya bint Al Hussein (born 3 May 1974) is the daughter of King Hussein of Jordan from his third wife, Queen Alia. Princess Haya is the junior wife of Sheikh Mohammed bin Rashid Al Maktoum. She is known as Her Royal Highness Princess Haya of Jordan , a title derived from her father. 5. Charlotte Casiraghi (born 3 August 1986) is the second child of Caroline, Princess of Hanover, and Stefano Casiraghi, an Italian industrialist. She is fifth in line to the throne of Monaco. See also: Celebrities with horses (50 photos) 4. Fawzia Fuad of Egypt (5 November 1921 – 2 July 2013) was an Egyptian princess who became Empress of Iran as the first wife of Shah Mohammad Reza Pahlavi. See also: The most beautiful black models 3. Princess Ameera Al-Taweel (born 6 November 1983) in the English-language press, is a Saudi Arabian princess and philanthropist and International fashion icon. See also: All winners of the contest Miss World Muslimah 2. Rania Al Abdullah (born 31 August 1970) is the Queen consort of Jordan. 1. Grace Kelly, HSH Princess Grace of Monaco (November 12, 1929 – September 14, 1982) was an American film actress and Princess of Monaco as the wife of Prince Rainier III. See also: The most beautiful Hollywood actresses\n\n\n \n\n \n\nПодробнее: <http://beauty-around.com/en/tops/item/878-most-beautiful-queen-and-princess-of-the-world>\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
496fceb0637232b71952692104c31ae3 | What are the best tutorial site to learn Greenplum Database and PostgreSQL? | tutorialdba.com | 2019.43 | [
{
"text": "\nLet me help you out with the Greenplum since I kicked off my career with Greenplum some 3.5 years back.\n\n\nIf you don't know anything about Greenplum, I recommend you to study with For Greenplum database tutorial : <https://2ndquadrant.in/courses/it-software/databases/greenplum-database/>\n\n\nThere you can find the introduction courses on various Pivotal products, help yourself to find Greenplum(GPDB) with title \"Introduction to Greenplum Database\".\n\n\nPostgreSQL as far as I'm concerned is popular for its fancy stored procedures other extended features. \n\n \n\n[The world's most advanced open source database](http://www.postgresql.org/) PostgreSQL org has all the tutorials required to work on PostgreSQL. \n\n \n\n......but for everything else you always have [www.tutorialdba.com](http://www.tutorialdba.com/) & <https://discuss.tutorialdba.com/postgresql-dba>. They have plenty good responses and solutions for possibly all problems you would facing working on it.\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
9a58fea49960b1228e7ce4c50790f2cb | What is the difference between nice and renice values in linux | tutorialdba.com | 2020.45 | [
{
"text": "\nthe nice command allows you to change the priority of a process (that isn’t already running) while the renice command allows you to change the priority of a process that’s already running\n\n\nnote: you cannot change the priority of a running process using nice\n\n\nboth take values from -10 to 19 and the lower the value the higher the priority\n\n\nthe default value for any running process is 0\n\n\nany user can increase the value of a process (slow it down) but it requires root access to decrease it (make it faster)\n\n\nand finally you have to know the PID of a process in order to change its priority using the renice command\n\n\n**An example of renice command is as follows:**\n\n\n\n```\n**renice -5 -p 3434** \n```\n\nthis will set the priority of process id no 3434 to -5, which will inturn increase its priority over others\n\n\n**An examples of the nice command is as follows:**\n\n\n\n```\nnice -n 11 <command name>\n```\n\n \n\n\n[How to Check Nice Value of Linux Processes](https://discuss.tutorialdba.com/783/check-nice-value-linux-processes)\n==================================================================================================================\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
b92adbd8add796b955b4c82652e00d04 |
Plus, on the red carpet, Meagher argues, these brands would be competing against the conglomerate backed heavyweights, with **[Off White Hoodie](https://www.offwhiteclothes.com/)** their big budgets and contracted talent street style gives them a chance to stand on their own. It's also an opportunity for a celebrity to support historically marginalized designers and give them a major PR boost, Smith notes From that visibility, maybe an investor will see them maybe press will see them and do a feature on them.
T shirts are like a blank slate, Mayock explained to us. As a category, they're the perfect vehicle for someone to voice what it's like to be **[Off White T Shirt](https://www.offwhiteclothes.com/)** in their own skin. Joining this family not just on screen, but off screen too. Our incredibly creative and generous cast, crew, outstanding fans it's all been beyond anything I could have imagined.
Transforming a local music and film festival into an internationally renowned event is no small feat, but Afropunk co founder Matthew Morgan is preparing to do it for the second time. September 15 marks the start of National Hispanic Heritage Month a time when many nations in Central America celebrate their national independence and Latinx and Hispanic people are honored for their history and contributions.
The line, which includes clothing, hats, scarves and accessories, are made with the intention of promoting awareness of Inuit and Northern culture. It's a different kind of heat here, she says. But there's another huge plus of summer coming to an end End of summer sales. They're happening all over the internet right now, but we're particularly keen on the one that's going on at Nordstrom because discounts are up to 60 percent off.
Sweaters, jackets and long sleeve blouses are just the beginning of the layering technique. Boots and booties provide fashionistas with the perfect opportunity to both be chic and stay warm. Come October 31, instead of digging deep into our style encyclopedias for a punny costume, we're turning the camera on ourselves and not just for a selfie. No, **[Off White Store](https://www.offwhiteclothes.com/)** we're not talking about a shoefie or an OOTD hell, not even an evachenpose.
| Off White Hoodie their big | tutorialdba.com | 2021.43 | [
{
"text": "\n[**Travis Scott Jordan 1**](https://www.travisscott-jordan1.com/) \n\n \n\n[**Air Jordans**](https://www.airsjordans.com/) \n\n \n\n[**Jordan 11s**](https://www.jordan-11s.com/) \n\n \n\n[**Jordan 11**](https://www.jordans-11.com/) \n\n \n\n[**Jordans Shoes**](https://www.jordansshoes.org/) \n\n \n\n[**Retro Jordans**](https://www.retro-jordan.com/) \n\n \n\n[**Moncler Jackets**](https://www.monclerjackets.us.com/) \n\n \n\n[**Nike Air Jordan**](https://www.nikeair-jordan.com/) \n\n \n\n[**Moncler Outlet**](https://www.moncler-outlets.com/) \n\n \n\n[**Off-White**](https://www.off-white.us.org/) \n\n \n\n[**Yeezy 450**](https://www.yeezy-450.com/) \n\n \n\n[**Yeezy 500**](https://www.yeezys500.com/) \n\n \n\n[**Yeezy**](https://www.yeezyyeezy.com/) \n\n \n\n[**Yeezy 700**](https://www.yeezys-700.com/) \n\n \n\n[**Yeezy Supply**](https://www.yeezys-supply.com/) \n\n \n\n[**Off White Shoes**](https://www.offwhiteshoess.com/) \n\n \n\n[**NFL Jerseys**](https://www.nflsjerseys.us.com/) \n\n \n\n[**Jordan Shoes**](https://www.jordanshoes.org/) \n\n \n\n[**Jordans Shoes**](https://www.jordans-shoes.com/) \n\n \n\n[**Yeezy 350 V2**](https://www.yeezy350-v2.com/) \n\n \n\n[**Adidas Yeezy**](https://www.adidasyeezys.com/) \n\n \n\n[**Yeezy**](https://www.yeezyoutlet.us.com/) \n\n \n\n[**Yeezy 700**](https://www.yeezy-700.us.com/) \n\n \n\n[**Yeezy**](https://www.yeezyv2.us.com/) \n\n \n\n[**Nike Outlet**](https://www.nike-outlets.com/) \n\n \n\n[**Yeezy Shoes**](https://www.yeezy-shoes.us.com/) \n\n \n\n[**UNC Jordan 1**](https://www.uncjordan1.us/) \n\n \n\n[**Jordan 13**](https://www.jordan-13.us/) \n\n \n\n[**Jordan AJ 1**](https://www.jordanaj1.com/) \n\n \n\n[**Yeezy Foam Runner**](https://www.yeezyfoam-runner.com/) \n\n \n\n[**Nike Outlet**](https://www.nikestoreoutlet.us.com/) \n\n \n\n[**Nike Outlet**](https://www.nikeoutletfactory.us/) \n\n \n\n[**AJ1**](https://www.aj1.us.com/) \n\n \n\n[**Yeezy Supply**](https://www.yeezy-supply.com/) \n\n \n\n[**Yeezy Zebra**](https://www.yeezy-zebra.com/) \n\n \n\n[**Jordan 1 Low**](https://www.jordan1low.com/) \n\n \n\n[**Air Jordans**](https://www.air-jordans.us.org/) \n\n \n\n[**Pandora Charms**](https://www.pandoracharms.uk.com/) \n\n \n\n[**Adidas UK**](https://www.adidasuk.uk.com/) \n\n \n\n[**Nike Store**](https://www.nikestoreoutlet.us.com/) \n\n \n\n[**Adidas Yeezy Official Website**](https://www.adidasyeezyofficialwebsite.com/) \n\n \n\n[**Yeezy 350**](https://www.yeezy350.us.com/) \n\n \n\n[**Jordan 1**](https://www.jordan1.uk.com/) \n\n \n\n[**Nike Outlet**](https://www.nikesoutlet.us.com/) \n\n \n\n[**YEEZY SUPPLY**](https://www.supplyyeezys.us/) \n\n \n\n[**Pandora Charms**](https://www.pandoracharms.cc/) \n\n \n\n[**Nike Shoes**](https://www.nikeshoes.cc/) \n\n \n\n[**Nike Outlet**](https://www.nikeoutlet.uk.com/) \n\n \n\n[**Pandora Outlet**](https://www.pandoraoutlet.org/) \n\n \n\n[**Jordan Shoes**](https://www.jordanshoess.com/) \n\n \n\n[**Air Jordan 4**](https://www.air-jordan4.com/) \n\n \n\n[**Pandora Jewelry**](https://www.pandorajewelryusa.us.com/) \n\n \n\n[**Pandora Rings**](https://www.ringspandora.com/) \n\n \n\n[**Pandora Bracelets**](https://www.bracelets-pandora.com/) \n\n \n\n[**Adidas Yeezy**](https://www.yeezy-adidas.us.com/) \n\n \n\n[**Yeezy**](https://www.yzy.us.com/) \n\n \n\n[**Pandora Charms**](https://www.charmspandora.us.com/) \n\n \n\n[**Nike Outlet**](https://www.nikeoutletstore.us.com/) \n\n \n\n[**Adidas Yeezy**](https://www.adidas-yeezy.org/) \n\n \n\n[**Air Max 720**](https://www.airmax-720.com/) \n\n \n\n[**Nike Air Max 270**](https://www.nike-airmax270.com/) \n\n \n\n[**Air Jordan 11**](https://www.air-jordan11.com/) \n\n \n\n[**Air Force 1**](https://www.air-force1.com/) \n\n \n\n[**Air Jordan 1**](https://www.air-jordan1.com/) \n\n \n\n[**Nike Jordans**](https://www.nike-jordans.com/) \n\n \n\n[**Jordan 1s**](https://www.jordan-1s.com/) \n\n \n\n[**Pandora UK**](https://www.pandorauk.uk.com/) \n\n \n\n[**Nike Jordan 1**](https://www.nikejordan1.com/) \n\n \n\n[**Jordan 1**](https://www.jordan-1.org/) \n\n \n\n[**Yeezy Slides**](https://www.yeezyslides.us.com/) \n\n \n\n[**Nike Air VaporMax**](https://www.nikeairvapormax.us/) \n\n \n\n[**Nike Vapormax Flyknit**](https://www.nikevapormaxflyknit.com/) \n\n \n\n[**Air Jordan 1 Mid**](https://www.airjordan1-mid.com/) \n\n \n\n[**Adidas yeezy**](https://www.yeezyadidas.de/) \n\n \n\n[**Yeezy Shoes**](https://www.yeezy-shoess.com/) \n\n \n\n[**Adidas Yeezy**](https://www.adidasyeezy.me.uk/) \n\n \n\n[**Yeezy 350**](https://www.yeezy350.de/) \n\n \n\n[**Nike Shoes**](https://www.nikes.us.com/) \n\n \n\n[**Nike Outlet**](https://www.nikeoutletstoreonlineshopping.us/) \n\n \n\n[**Yeezy**](https://www.yeezystore.us.com/) \n\n \n\n[**NFL Shop Official Online Store**](https://www.nflshopofficialonlinestore.com/) \n\n \n\n[**Nike UK**](https://www.nikeuk.uk.com/) \n\n \n\n[**Yeezy**](https://www.yeezy.uk.com/) \n\n \n\n[**Yeezy 350**](https://www.yeezy350.uk.com/)\n\n\n",
"name": "",
"is_accepted": false
}
] |
1198141032275bc3f07f9333a027a68b |
[Sales Incentive Programs](http://www.level6incentives.com/salesperson-incentives/) | What is a Spiff- Level 6 Incentives: What is a spiff? Spiff (Sales Program Incentive Funds) is one of the sales incentive programs used by businesses and companies to drive sales.
For more information about Sales Incentive Programs visit website level6incentives.com
| Sales Incentive Programs | tutorialdba.com | 2021.43 | [
{
"text": "\nUnited Airlines is a main booked flight specialist organization as of now. Subsequently, they offer all that could be within reach to ensure that their travelers are cheerful. What's more, with regards to United Airlines cancellations, you can't miss the different techniques presented by them. Continue reading for more : [united airlines cancel flight](https://www.bookunitedairlines.com/united-airlines-cancellation-policy/)\n\n\n",
"name": "",
"is_accepted": false
}
] |
434e2c5993fe18c5ef97e2777e94c044 | How to migrate the database from mssql to postgresql ? | tutorialdba.com | 2020.34 | [
{
"text": "\nIspirer MnMTK converts database schema (tables and SQL objects), business logic and data <https://www.ispirer.com/products/sql-server-to-postgresql-migration> \n\n\ncheck if it suits\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
616354d27b1127737160aa5e5cd43b0d | What is PostgreSQL Json Data Type ? And how to handle & perform insert,up upda, delete | What is PostgreSQL Json Data Type | tutorialdba.com | 2021.04 | [
{
"text": "\n**what is JSON ?**\n\n\n* It is one of the data type in postgres.\n* Json + PostgreSQL =NoSQL\n* JSON STANDS FOR JavaScript Object Notation\n* JSON data types stored value is valid according to the JSON rules.\n* JSON data types are for storing JSON data as multi-level, dynamically structured object graphs.\n* serialised object is stored in a text column. The json type takes care of deserialising it back to object graph while reading values from that column.\n* The main usage of JSON is to transport data between a server and web application. Unlike other formats, JSON is human-readable text.\n* PostgreSQL supports JSON data type since version 9.2.\n* It provides many functions and operators for manipulating JSON data.\n\n\n**There are two JSON data types:** \n\n1.json \n\n2.jsonb\n\n\n* The json data type stores an exact copy of the input text.\n* jsonb data is stored in a decomposed binary format\n* jsonb Insertion makes it slightly slower to input due to added conversion overhead.\n* jsonb also supports indexing\n* JSONB does not keep duplicate object keys. If duplicate keys are specified in the input, only the last value is kept.\n\n\n**PRACTICAL 1**. CREATING SIMPLE JSON TABLE: \n\n \n\n\n\n```\npostgres=# CREATE TABLE SALES (\n ID INT NOT NULL PRIMARY KEY,\n SALES\\_INFO json NOT NULL\n);\n```\n\n**From above Sales table consists of two columns:**\n\n\n* The id column is the primary key column that identifies the sales id.\n* The sales\\_info column stores the data in the form of JSON data types.\n\n\n--Describe the sales table using \\d \n\n \n\n\n\n```\npostgres=# \\d sales\n Table \"public.sales\"\n Column | Type | Modifiers \n------------+---------+-----------\n id | integer | not null\n sales\\_info | json | not null\nIndexes:\n \"sales\\_pkey\" PRIMARY KEY, btree (id) \n```\n\n**PRACTICAL 2**. INSERTING JSON DATA ON SALES TABLE: \n\n \n\n\n\n```\nINSERT INTO SALES VALUES\n (1,'{ \"customer\": \"NIJAM\", \"PRODUCTS\": {\"product\\_name\": \"choclate\",\"total\\_item\": 6}}'\n );\n\nINSERT INTO SALES VALUES\n (2,'{ \"customer\": \"ABU\", \"PRODUCTS\": {\"product\\_name\": \"badam\",\"total\\_item\": 5}}'\n );\n\nINSERT INTO SALES VALUES\n (3,'{ \"customer\": \"UMAR\", \"PRODUCTS\": {\"product\\_name\": \"mobile\",\"total\\_item\": 1}}'\n ); \n```\n\n--List down the sales table \n\n \n\n\n\n```\npostgres=# select * from sales;\n id | sales\\_info \n----+----------------------------------------------------------------------------------\n 1 | { \"customer\": \"NIJAM\", \"PRODUCTS\": {\"product\\_name\": \"choclate\",\"total\\_item\": 6}}\n 2 | { \"customer\": \"ABU\", \"PRODUCTS\": {\"product\\_name\": \"badam\",\"total\\_item\": 5}}\n 3 | { \"customer\": \"UMAR\", \"PRODUCTS\": {\"product\\_name\": \"mobile\",\"total\\_item\": 1}}\n(3 rows) \n```\n\n--list down the json column only \n\n \n\n\n\n```\npostgres=# select sales\\_info from sales;\n sales\\_info \n----------------------------------------------------------------------------------\n { \"customer\": \"NIJAM\", \"PRODUCTS\": {\"product\\_name\": \"choclate\",\"total\\_item\": 6}}\n { \"customer\": \"ABU\", \"PRODUCTS\": {\"product\\_name\": \"badam\",\"total\\_item\": 5}}\n { \"customer\": \"UMAR\", \"PRODUCTS\": {\"product\\_name\": \"mobile\",\"total\\_item\": 1}}\n(3 rows) \n```\n\n--List down the id column only for understanding purposes. \n\n \n\n\n\n```\npostgres=# select id from sales;\n id \n----\n 1\n 2\n 3\n(3 rows) \n```\n\n**Deleting json objects:** \n\n \n\n\n\n```\npostgres=# select * from sales;\n id | sales\\_info \n----+-------------------------------------------------------------------------------\n 2 | { \"customer\": \"ABU\", \"PRODUCTS\": {\"product\\_name\": \"badam\",\"total\\_item\": 5}}\n 3 | { \"customer\": \"UMAR\", \"PRODUCTS\": {\"product\\_name\": \"mobile\",\"total\\_item\": 1}}\n 4 | { \"customer\": \"junaith\", \"PRODUCTS\": {\"product\\_name\": \"pen\",\"total\\_item\": 8}}\n 7 | { \"customer\": \"daniel\", \"PRODUCTS\": {\"product\\_name\": \"car\",\"total\\_item\": 8}}\n 8 | { \"customer\": \"daniel\", \"PRODUCTS\": {\"product\\_name\": \"car\",\"total\\_item\": 8}}\n(5 rows)\n\npostgres=# delete from sales where sales\\_info ->'PRODUCTS'->>'total\\_item'='8';\nDELETE 3\npostgres=# select * from sales; \n id | sales\\_info \n----+-------------------------------------------------------------------------------\n 2 | { \"customer\": \"ABU\", \"PRODUCTS\": {\"product\\_name\": \"badam\",\"total\\_item\": 5}}\n 3 | { \"customer\": \"UMAR\", \"PRODUCTS\": {\"product\\_name\": \"mobile\",\"total\\_item\": 1}}\n(2 rows) \n```\n\n \n\n\n**PRACTICAL 3**.HOW TO QUERYING & FILTERING JSON DATA: \n\nthe two operators -> and ->> to help you query JSON data.\n\n\n* -> will return the attribute as a JSON object key(original JSON type).\n* ->> will return the property as integer or text (the parsed form of the attribute).\n\n\n--See the below example first two query did'nt specify column name that is why it showing ?column? after that i specified the column name as customer\\_name \n\n \n\n\n\n```\npostgres=# select sales\\_info -> 'customer' from sales;\n ?column? \n----------\n \"NIJAM\"\n \"ABU\"\n \"UMAR\"\n(3 rows)\n\npostgres=# select sales\\_info ->> 'customer' from sales;\n ?column? \n----------\n NIJAM\n ABU\n UMAR\n(3 rows)\n\npostgres=# select sales\\_info ->> 'customer' as customer\\_name from sales;\n customer\\_name \n---------------\n NIJAM\n ABU\n UMAR\n(3 rows)\n\npostgres=# select sales\\_info -> 'PRODUCTS' from sales; \n ?column? \n----------------------------------------------\n {\"product\\_name\": \"choclate\",\"total\\_item\": 6}\n {\"product\\_name\": \"badam\",\"total\\_item\": 5}\n {\"product\\_name\": \"mobile\",\"total\\_item\": 1} \n```\n\n* Using -> operator returns a JSON object, you can chain it with the operator ->> to retrieve a specific node.\n* JSON IS CASESENSITIVE DEFAULTY, from below example first i try to retrieve the data using \"products\"(lowercase) that is why it showed empty value after that i chaneged \"PRODUCTS\"(UPPER CASE) then it showing values.\n\n\n\n```\npostgres=# select sales\\_info -> 'products' ->>'product\\_name' from sales;\n ?column? \n----------\n \n \n \n(3 rows)\n\n\npostgres=# select sales\\_info -> 'PRODUCTS' ->'product\\_name' from sales;\n ?column? \n------------\n \"choclate\"\n \"badam\"\n \"mobile\"\n(3 rows)\n\npostgres=# select sales\\_info -> 'PRODUCTS' ->>'product\\_name' from sales;\n ?column? \n----------\n choclate\n badam\n mobile\n(3 rows \n```\n\nFROM above example First sales\\_info -> 'PRODUCTS' returns as JSON objects. And then sales\\_info -> 'PRODUCTS' ->>'product\\_name' returns all products as text.\n\n\nMore information https[://www.tutorialdba.com/p/postgresql-json-data-type.html?m=1](https://www.tutorialdba.com/p/postgresql-json-data-type.html?m=1)\n\n\n",
"name": "",
"is_accepted": true
}
] |
8d01559db26238e53087cff644f5b6cf |
### Tomorrow's Prediction Marksans Pharma
| What is the target price of marksans pharma in 2021 june 5 | tutorialdba.com | 2021.25 | [
{
"text": "\nTomorrow's Prediction Marksans Pharma (MARKSANS) share price targets 74 rupees\n==============================================================================\n\n\n",
"name": "",
"is_accepted": false
}
] |
753df170eb47c5ade44bbd1e08097590 | What is the prerequest for Database creation and tablespace Creation as well as postgresql restoration to new server or database refreshment . | How to create Postgresql database in real time production environment ? | tutorialdba.com | 2020.40 | [
{
"text": "\nBefore creating database you have to create tablespace and then you have to give right tablespace name for database Creation \n\n\nAnd before creating tablespace you have to check with partion or disk space availability it's a prerequest for Postgresql tablespace creation\n\n\nHere given Following steps for creating tablespace & Database in Postgresql environment system.\n\n\nYou have to follow given steps only for Postgresql database creation & tablespace Creation.\n\n\nHere am performed simple steps of Postgresql Database restoring to some other new server & before that what prerequest want to follow for the postgresql Database restoring & Everything am mentioned here.\n\n\n**Checking the postgreSQL server status Before restoring :**\n\n\n\n```\ncd C:\\Program Files\\edb\\enterprincedb\\as9.6\\bin pg\\_ctl -D \"C:\\Program Files\\edb\\enterprincedb\\as9.6\\data\" status\n```\n\n**Connecting the database :**\n\n\n\n```\ncd C:\\Program Files\\edb\\enterprincedb\\as9.6\\bin\npsql -U enterprisedb -d edb\n```\n\n**Checking the tablespace & database whether already exist with same name in our server :** \n\nFOR TABLESPACE :-\n\n\n\n```\nSELECT spcname FROM pg\\_tablespace;\n```\n\nThe psql program's **\\db** meta-command is also useful for listing the existing tablespaces.\n\n\nFOR DATABASE :-\n\n\n\n```\nSELECT datname FROM pg\\_database;\n```\n\nThe psql program's **\\l** meta-command is also useful for listing the existing databse.\n\n\n**Creating tablespace :**\n\n\n\n```\nCREATE TABLESPACE \"HealthCraft\\_LC\" LOCATION 'D:\\HealthCraft\\_LC';\nCREATE TABLESPACE \"HealthCraft\\_DC\" LOCATION 'D:\\HealthCraft\\_DC';\n```\n\nBefore creating tablespace you have to check the server available space and available partition this is one of the prerequest as well.\n\n\n**Creating Database :**\n\n\n\n```\nCREATE DATABASE \"HealthCraft\\_LC\" TABLESPACE \"HealthCraft\\_LC\";\nCREATE DATABASE \"HealthCraft\\_DC\" TABLESPACE \"HealthCraft\\_DC\";\n\n```\n\nBefore Restoring Database you have to restore the roles other wise you will get roles does not exist error\n\n\n**Taking postgres Roles only backup :**\n\n\n\n```\npg\\_dumpall -h localhost -p 5432 U enterprisedb -v --roles-only -f \"/path/to/Preprod\\_roles\\_01012019.sql\"\n\n```\n\n**Taking Databases backup :**\n\n\n\n```\npg\\_dump -d HealthCraft\\_LC -U enterprisedb -f \"D:\\HealthCraft\\_LC\\_preprod\\_01012019.sql\"\npg\\_dump -d HealthCraft\\_DC -U enterprisedb -f \"D:\\HealthCraft\\_DC\\_preprod\\_01012019.sql\"\n```\n\n**Restoring the database roles it is the prerequest :**\n\n\n\n```\npsql -U enterprisedb -d edb -f \"D:\\tutorialdba\\Preprod\\_roles\\_01012019.sql\"\n```\n\nRestoring the postgresql database dump by using **psql** utility otherwise if your dump is custom format means you have to use **pg\\_restore** utility for database restoration :\n\n\n\n```\npsql -U enterprisedb -d \"HealthCraft\\_LC\" -f \"D:\\tutorialdba\\HealthCraft\\_LC\\_preprod\\_01012019.sql\"\npsql -U enterprisedb -d \"HealthCraft\\_DC\" -f \"D:\\tutorialdba\\HealthCraft\\_DC\\_preprod\\_01012019.sql\"\n```\n\n**else you can use \\i at postgres SQL console (if dump is plain format):**\n\n\n\n```\npsql -U enterprisedb -d HealthCraft\\_LC\n\\i D:\\HealthCraft\\_LC\\_preprod\\_01012019.sql\n```\n\nConnect the other(HealthCraft\\_DC) DB the restore the appropriate dump of HealthCraft\\_DC database\n\n\n\n```\n\\c HealthCraft\\_DC\n\\i D:\\HealthCraft\\_DC\\_preprod\\_01012019.sql\n```\n\n**To Dropping the postgres database :**\n\n\n\n```\ndrop DATABASE \"HealthCraft\\_LC\";\nERROR: database \"HealthCraft\\_LC\" is being accessed by other users\nDETAIL: There are 2 other sessions using the database.\n```\n\nFirst of all you have to kill the connected session by using **pg\\_terminate\\_backend**\n\n\n\n```\nSELECT pg\\_terminate\\_backend(pg\\_stat\\_activity.pid) FROM pg\\_stat\\_activity WHERE pg\\_stat\\_activity.datname = 'HealthCraft\\_LC' AND pid <> pg\\_backend\\_pid();\npg\\_terminate\\_backend\n----------------------\nt\nt\n(2 rows)\n```\n\nafter kill the session you can drop the database by using drop command\n\n\n\n```\ndrop DATABASE \"HealthCraft\\_LC\";\n```\n\nSome time session will be connected automatically again and again at the time you have to issue both command without time delay.\n\n\n**for example**\n\n\n\n```\nSELECT pg\\_terminate\\_backend(pg\\_stat\\_activity.pid) FROM pg\\_stat\\_activity WHERE pg\\_stat\\_activity.datname = 'HealthCraft\\_LC' AND pid <> pg\\_backend\\_pid();\ndrop DATABASE \"HealthCraft\\_LC\";\n## copy this above two lines paste it to sql prompt\n```\n\n**Running the script at background**\n\n\nWindows Server :\n\n\n\n```\nSTART \"\" psql -U enterprisedb -d \"HealthCraft\\_LC\" -f \"D:\\HealthCraft\\_LC\\_preprod\\_01012019.sql\" \n```\n\nLinux Server :\n\n\n\n```\n ./psql -U enterprisedb -d \"HealthCraft\\_LC\" -f D:\\HealthCraft\\_LC\\_preprod\\_01012019.sql &\n```\n\n",
"name": "",
"is_accepted": false
}
] |
f8179f93a750273db5985e976b633111 | Can anyone tell about the scan listener and scan IP in short | tutorialdba.com | 2021.04 | [
{
"text": "Scan ip address is just a IP address scan listener is the listener. If cluster create the scan listener the scan ip adress must be there without scan ip scan listener can not be created \n\n \n\nAt the time of grid infrastructure installation \n\n \n\nScan-ip & scan-listener forms a pair",
"name": "",
"is_accepted": false
}
] |
|
3d5ba5662127aa8bb42deb1228d2bcd8 | Hi guys Pls. Let me know how to check count of all procedures | tutorialdba.com | 2019.43 | [
{
"text": "Select count(*) from pg\\_proc;",
"name": "",
"is_accepted": false
}
] |
|
e68eb97c7ad39a019980341c8f5cb971 |
how to optimize and repair *MySQL* databases and monitoring of your *MySQL* server including CPU/RAM
| MySQL Maintenance Tasks | tutorialdba.com | 2020.45 | [
{
"text": "\nFrom time to time, MySQL 5.1 databases need a little house keeping. We found our production DB had a hard time running a simple join query between two tables with about 400k rows. It was taking between 30 and 100 seconds to run. On QA however, it was taking 58 milliseconds. The columns involved were already indexed. Thankfully it wasn’t impacting our users, but it still bugged me. The solution was simple, just run some cleanup commands. After the cleanup, on the live server the same query took just 4.8 milliseconds – that’s more like it!\n\n\n**Summary of solution:**\n\n\n1. Backup database\n2. Check\n3. Optimize\n4. Analyze\n\n\n\n```\n$ mysqldump -u root -p --create-options --routines --triggers dbname > ./db.dmp\n# note these cause LOCKS, so be careful on your production server!\n$ mysqlcheck -u root -p --check --databases dbname\n$ mysqlcheck -u root -p --optimize --databases dbname\n$ mysqlcheck -u root -p --analyze --databases dbname\n```\n\n \n\n\n**Complete details about each step:**\n\n\n**1) First make database backup with mysqldump:** \n\nDon’t forget the argument *–routines* if you have stored procedures or functions and *–triggers* if using triggers:\n\n\n\n```\n$ mysqldump -u root -p --create-options --routines --triggers dbname > ./db.dmp\n# copy to another server\n$ scp ./db.dmp user@somehost:~/\n```\n\nFor bonus points, *actually restore the database on another system* to make sure you have a valid backup.\n\n\nThis step may be impractical if the database is huge. In that case you are probably already using replication and have a backup system worked out.\n\n\n**2) Check:** \n\nChecks table for integrity errors. \n\n<http://dev.mysql.com/doc/refman/5.1/en/check-table.html>\n\n\nTo check a single table:\n\n\n\n```\nmysql> CHECK TABLE {table name};\n```\n\nTo check all tables in a database, from command line:\n\n\n\n```\n$ mysqlcheck -u root -p --check --databases dbname\n```\n\nThis seems like a really smart thing to do on a regular basis.\n\n\n**3) Optimize:**\n\n\nLike a defrag operation, the optimize tables command reclaims unused space. At least, that is what it does for MyISAM. With InnoDB it basically runs an ALTER TABLE statement that changes nothing but tells MySQL to rebuild the table and its indexes. \n\n<http://dev.mysql.com/doc/refman/5.1/en/optimize-table.html>\n\n\nTo optimize a single table:\n\n\n\n```\nmysql> OPTIMIZE TABLE {table name};\n```\n\nTo optimize all tables in a database, from command line:\n\n\n\n```\n$ mysqlcheck -u root -p --optimize --databases dbname\n```\n\nIf you get “Table does not support optimize, doing recreate + analyze instead”, that is normal for InnoDB.\n\n\n**4) Analyze:**\n\n\nAnalyze rebuilds and optimizes the performance of indexes, specially it rebuilds the key distribution. If you have a slow running query but indexes are in place, it might be time to run this. A read lock goes into effect while this is running. If you have only InnoDB tables, this is already taken care of by Optimize.\n\n\n<http://dev.mysql.com/doc/refman/5.1/en/analyze-table.html>\n\n\nTo analyze a single table:\n\n\n\n```\nmysql> ANALYZE TABLE {table name};\n```\n\nTo analyze all tables in a database, from command line:\n\n\n\n```\n$ mysqlcheck -u root -p --analyze --databases dbname\n```\n\nWith InnoDB and ANALYZE TABLE, there are some oddities. In particular, the number of samples the analyzer takes can vary (configuration option is innodb\\_stats\\_sample\\_pages). The default is low, and this means running analyze tables repeatedly will produce slightly different results.\n\n\n",
"name": "",
"is_accepted": true
},
{
"text": "\nMail alert notification MariaDB maintainence task activity\n\n\n\n```\nEMAIL=\"[email protected]\"\n\nmysqlcheck -c -A\nmysqlcheck -o -A\nmysqlcheck -a -A\nmysqlcheck --auto-repair -A\n\n##SEND AN EMAIL###\n MESSAGE=\"$(tail -15 /var/lib/publishman.com.err)\"\n SUBJECT=\" Maintenance Done on $(date) \"\n echo -e \" LOGS : \\n$MESSAGE \" | mail -s \"$SUBJECT\" \"$EMAIL\"\n```\n\n",
"name": "",
"is_accepted": false
},
{
"text": "\nPerform Routine Table Checks, Optimization, and Repairs\n-------------------------------------------------------\n\n\nTables should be actively maintained and checked regularly as a proactive measure of problem prevention. When problems do arise, this step will help minimize the impact of the problem on the server and data. The mysqlcheck binary program can be used to perform most of these checks. The mysqlcheck program can be used to perform database table consistency checks, table optimization, table analysis, or table repair by passing different arguments to the CLI.\n\n\nTIP: you can create renamed binary copies if the mysqlcheck OR use symbolic links to make calls to the CLI more intuitive. This is because there is a special feature in the code that permits the default behavior of the CLI to change when the file or link is named in a certain fashion. \n\n\nYou can achieve this behavior by:\n\n\n\n```\ncp mysqlcheck mysqlrepair\ncp mysqlcheck mysqloptimize\ncp mysqlcheck mysqlanalyze\n```\n\nOR\n\n\n\n```\nln -s mysqlcheck mysqlrepair\nln -s mysqlcheck mysqloptimize\nln -s mysqlcheck mysqlanalyze\n```\n\nIt is reccommended that this CLI be added to the crontab of the UNIX user account that mysql runs under. \n\n\nCreate a ~/.my.cnf file with login credentials so that the MySQL CLIs can run non-interactively in cron and be able to authenticate.\n\n\n\n```\n[client]\nuser=myuser\npassword=myuserpassword\n```\n\nProtect the file by tightening the UNIX file permissions so that only the UNIX user and/or root may view the file, since it has a password in clear text.\n\n\n\n```\nchmod 600 ~/.my.cnf\nls -lta .my.cnf\n-rw------- 1 mysql mysql 37 2008-12-14 12:01 .my.cnf\n```\n\nAdd the following to the crontab of the UNIX user (in most cases mysql) that MySQL runs under:\n\n\n\n```\n0 0 * * 0 /usr/local/mysql/bin/mysqlcheck --all-databases --check-only-changed --silent\n```\n\nWhen a table is being checked by mysqlcheck, it cannot be updated. So, performing mysqlcheck on large tables should be used at your discretion. It may not make sense to perform mysqlcheck on large tables if you cannot afford the table to be unavailable for updates.\n\n\nAlternatively, you could create the .my.cnf in the /root directory on Linux systems, and place the job into the /etc/cron.daily directory. On Fedora 10, I placed the .my.cnf file in the /root directory. An executable shell script with the mysqlcheck command was placed in the /etc/cron.daily directory. The /etc/crontab script on linux calls a run-parts script which runs all scripts it finds in the /etc/cron.daily, /etc/cron.hourly, /etc/cron.weekly, and /etc/cron.monthly directories.\n\n\n",
"name": "",
"is_accepted": false
}
] |
59073f0e45b2069618711e44d8a11f74 | How to increase the priority of linux group's process ? | tutorialdba.com | 2020.45 | [
{
"text": "\nYou can even change the priorities of all running group belongs to a particular group like below.\n\n\n\n```\n# renice -n -20 -g nijamgroup\n```\n\nThe above command changes the priority of all running processes which are belongs to a group called nijamgroup.\n\n\n",
"name": "",
"is_accepted": true
},
{
"text": "renice -n value -g groupname",
"name": "",
"is_accepted": false
}
] |
|
ebfee9a8913318fc1bfb7909d80e4216 | What is the current status of Corona virus ? | tutorialdba.com | 2020.34 | [
{
"text": "\nThe number of Americans under virtual lockdown grew Saturday to over 80 million, with New Jersey joining the list of states issuing a stay-at-home order or other sweeping mandates to fight the spread of the new [coronavirus](https://www.nbcnews.com/health/health-news/live-blog/coronavirus-updates-california-issues-stay-home-order-no-new-local-n1164541).\n\n\n“We must flatten the curve and ensure residents are practicing social distancing,\" New Jersey Gov. Phil Murphy said, adding, “Even with this order in effect … life in New Jersey does not have to come to a complete standstill.”\n\n\nNew Jersey's stay-at-home order applies to nearly all of its 9 million residents. It comes after Illinois issued a similar mandate Friday for its 13 million residents and following California's stay-at-home order for its population of 40 million.\n\n\nIn New York, with a population of over 19 million, the governor has ordered that all nonessential businesses keep their workers home. [Pennsylvania's governor has also ordered](https://www.nbcphiladelphia.com/news/coronavirus/pa-governor-all-businesses-not-life-sustaining-must-shut-down/2333625/) that all businesses that are not \"life-sustaining\" close.\n\n\nNew York also announced Saturday that the number of coronavirus cases there now tops 10,000.\n\n\nWorldwide, there are now more than 275,000 confirmed cases of the coronavirus, according to [Johns Hopkins University](https://coronavirus.jhu.edu/map.html).\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
8680dcbd3313217c9e319ba450d3be8c | Shell script to watch the disk space mail alert | tutorialdba.com | 2020.16 | [
{
"text": "\ndf displays the amount of disk space available on the file system containing each file name argument. If no file name is given, the space available on all currently mounted file systems is shown. Read man page of df if you are new to df command.\n\n\n \n\n\n \n\n\n### Steps\n\n\n=> Find disk space using df\n\n\n=> Filter out filesystem and find out the percentage of space using grep\n\n\n=> Write a shell script\n\n\n### Step # 1: First get disk space:\n\n\n**$ df -H**\n\n\nOutput:\n\n\n\n```\nFilesystem Size Used Avail Use% Mounted on\n/dev/hdb1 20G 14G 5.5G 71% /\ntmpfs 394M 4.1k 394M 1% /dev/shm\n/dev/hdb5 29G 27G 654M 98% /nas/www\n\n```\n\n### Step # 2: Next filter out filesystem and find out the percentage of space\n\n\n`$ df -H | grep -vE '^Filesystem|tmpfs|cdrom' | awk '{ print $5 \" \" $1 }'`\n\n\nOutput:\n\n\n\n```\n71% /dev/hdb1\n98% /dev/hdb5\n\n```\n\n### Step # 3: Write a shell script\n\n\nAbove command displays field 5 and 1 of df command. Now all you need to do is write a script to see if the percentage of space is >= 90% (download [script](https://bash.cyberciti.biz/monitoring/monitor-disk-space-alert.bash.php)):\n\n\n\n```\n#!/bin/sh\ndf -H | grep -vE '^Filesystem|tmpfs|cdrom' | awk '{ print $5 \" \" $1 }' | while read output;\ndo\n echo $output\n usep=$(echo $output | awk '{ print $1}' | cut -d'%' -f1 )\n partition=$(echo $output | awk '{ print $2 }' )\n if [ $usep -ge 90 ]; then\n echo \"Running out of space \\\"$partition ($usep%)\\\" on $(hostname) as on $(date)\" |\n mail -s \"Alert: Almost out of disk space $usep%\" [email protected]\n fi\ndone\n\n```\n\n### Setup Cron job\n\n\nSave and install script as [cronjob](https://www.cyberciti.biz/faq/how-do-i-add-jobs-to-cron-under-linux-or-unix-oses/). Copy script to /etc/cron.daily/ (script [downolad link](https://bash.cyberciti.biz/monitoring/monitor-disk-space-alert.bash.php)) \n\n`# cp diskAlert /etc/cron.daily/ \n\n# chmod +x /etc/cron.daily/diskAlert`\n\n\nOR install as cronjob: \n\n`crontab -e`\n\n\nWrite cronjob as per your requirement \n\n`10 0 * * * /path/to/diskAlert`\n\n\n### Updated script version\n\n\nTony contributed and updated my script – You can exclude selected filesystem in case you don’t want monitor all filesystems.\n\n\n\n```\n#!/bin/sh\n# set -x\n# Shell script to monitor or watch the disk space\n# It will send an email to $ADMIN, if the (free available) percentage of space is >= 90%.\n# -------------------------------------------------------------------------\n# Set admin email so that you can get email.\nADMIN=\"root\"\n# set alert level 90% is default\nALERT=90\n# Exclude list of unwanted monitoring, if several partions then use \"|\" to separate the partitions.\n# An example: EXCLUDE\\_LIST=\"/dev/hdd1|/dev/hdc5\"\nEXCLUDE\\_LIST=\"/auto/ripper\"\n#\n#::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::\n#\nfunction main\\_prog() {\nwhile read output;\ndo\n#echo $output\n usep=$(echo $output | awk '{ print $1}' | cut -d'%' -f1)\n partition=$(echo $output | awk '{print $2}')\n if [ $usep -ge $ALERT ] ; then\n echo \"Running out of space \\\"$partition ($usep%)\\\" on server $(hostname), $(date)\" | \\\n mail -s \"Alert: Almost out of disk space $usep%\" $ADMIN\n fi\ndone\n}\n\nif [ \"$EXCLUDE\\_LIST\" != \"\" ] ; then\n df -H | grep -vE \"^Filesystem|tmpfs|cdrom|${EXCLUDE\\_LIST}\" | awk '{print $5 \" \" $6}' | main\\_prog\nelse\n df -H | grep -vE \"^Filesystem|tmpfs|cdrom\" | awk '{print $5 \" \" $6}' | main\\_prog\nfi\n\n```\n\n \n\n\n",
"name": "",
"is_accepted": false
}
] |
|
84a836525736decad3460cedb9c3f8d4 |
How to redirect my custom domain blogger ? after get SSL or without blogger SSL
Site link works : <https://www.tutorialdba.com>
But this does not work : <https://tutorialdba.com/>
Is there a solution to this problem ?
| How to redirect my custom domain blogger ? | tutorialdba.com | 2021.43 | [
{
"text": "\n**Go to your Blogger** Dashboard > Settings > Basic \n \nand Under Blog Address Click on Edit. \nNow Tick Redirect tutorialdba.com to www.tutorialdba.com \n \nWait for sometimes and You're Done.\n\n\n",
"name": "",
"is_accepted": true
}
] |
5af61e1d2f5046629ddb8ad664fafc94 | Hello All - what is the best way to change replica set name without downtime or minimal downtime. It is the 5 node replica set. | what is the best way to change replica set name without downtime | tutorialdba.com | 2019.43 | [
{
"text": "1. Stop the mongod service on all the 5 nodes \n\n2. Start 1 node without any replica set name in the config file and drop local DB and stop the service \n\n3. Change the config file back with the new replica set name in all the 5 nodes \n\n4. Start all the 5 servers with the new config file \n\n5. rs.initiate() node 1 and rs.add all rest 4 nodes",
"name": "",
"is_accepted": false
}
] |
53841551d5a00f604e351890c86dfa7e | How a Bengaluru software engineer caught coronavirus infection ? Covid-19 | tutorialdba.com | 2020.34 | [
{
"text": "\n* The techie had gone to Dubai from Bengaluru last month where he worked with people from Hong Kong\n* The Karnataka government is now keeping a watch on all those who came in contact with the patient in Bengaluru\n\n\n**BENGALURU** : The software engineer from Bengaluru, who become India's fifth confirmed case of coronavirus, had gone to Dubai last month where he is believed to have caught the infection. The 24-year-old techie, whose identity has been kept confidential, is employed by a software company in Bengaluru and had worked with people from Hong Kong in Dubai last month where he is suspected to have contracted the virus.\n\n\nAfter finishing work, the youngster returned to Bengaluru on February 20 and till then he showed no signs of any illness or infection. Two days later, he took a bus to Hyderabad, his native city, where he fell sick.\n\n\nThe techie took treatment for fever after coming to Hyderabad and after five days, he was admitted to a private super specialty hospital in the city. As his flu and fever like symptoms did not subside, he was shifted to the state-run Gandhi Hospital on Sunday evening, Telangana health minister E Rajender said.\n\n\n \n\n\nOn Monday, he tested positive for the novel [coronavirus](https://www.livemint.com/news/india/coronavirus-update-two-cases-reported-from-delhi-telangana-11583139808115.html)(COVID-19) and became Telangana's first case of the disease that emanated from China and has so far killed more than 3,000 people in several parts of the world.\n\n\nKarnataka Health Minister B Sriramulu has said that a medical team is monitoring the health condition of all those people who had stayed with the coronavirus-hit techie.\n\n\n\"It has come to our knowledge that the coronavirus-hit person in Hyderabad had gone from Bengaluru. Therefore, all the members in the house where he had stayed here have been identified and are under watch,\" Sriramulu tweeted on late Monday night.\n\n\nBefore moving to Hyderabad, the techie spent two days in Bengaluru, meeting company employees and other people. About 80 people, including eight family members, 42 staffers at the private hospital including doctors, paramedical staffers and front office personnel and 27 bus passengers, are believed to have come into contact with him since his return.\n\n\nThe family members of the techie were brought to Gandhi Hospital and kept in an isolation ward. The staffers at the private hospitals were also advised to undergo all tests.\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
b22dca89808a3d3e8ec2fe13e434b8fe | If we have multiple slaves under the master which went down and we have promoted a stand by slave to master. Then how to point all these slaves again under the new master ?
What about the data consistency part ? Means how to double sure that we don't have any kind of data loss while this switching ? | Multiple slave failover/switchover in postgres | tutorialdba.com | 2021.04 | [
{
"text": "Switching is an instant process. Moreover you will do promote only when your master is down and after promote you will need to make changes and point your application to new ip (new master) so data should be consistent and there should not be any data loss, also it depending upon in which situation your master went down and how it went down \n\n \n\nI guess that is only possible if you have a cascading setup, as per your current scenario your master had multiple slaves so if you promote any of slave and make it master I don't think there's a chance to make other servers as a standby on new master, \n\nFirst issue will be with timelien as new master will have a new time line \n\nAlso on all the old standby servers all the details are of old master, and I am not sure that changing recovery file will resume replication with new master",
"name": "",
"is_accepted": false
}
] |
21af757a55cacaff37fc6e2fa9369358 | How to check table if the table needs analyze on postgres? | tutorialdba.com | 2020.16 | [
{
"text": "\nIf table having large number of dead tuble you have to vacuum or analyze the table\n\n\n**For Example:**\n\n\n**STEP 1**. update the k1 table again for making operation of VACUUM (VERBOSE, ANALYZE)\n\n\n\n```\nupdate k1 set tableowner='john';\n\n```\n\n**STEP 2.**Check the dead tubles\n\n\n\n```\npostgres=# select n\\_dead\\_tup ,last\\_vacuum,last\\_analyze,n\\_tup\\_upd, n\\_tup\\_del,n\\_tup\\_hot\\_upd,relname ,seq\\_scan,idx\\_scan from pg\\_stat\\_all\\_tables where relname='k1';\n n\\_dead\\_tup | last\\_vacuum | last\\_analyze | n\\_tup\\_upd | n\\_tup\\_del | n\\_tup\\_hot\\_upd | relname | seq\\_scan | idx\\_scan \n------------+-------------------------------+--------------+-----------+-----------+---------------+---------+----------+----------\n 1880064 | 2017-05-15 05:20:05.216778-04 | | 4708352 | 8192 | 8333 | k1 | 21 | \n(1 row)\n\n```\n\n**STEP 3.**Now let us making VACUUM\n\n\n\n```\nVACUUM (VERBOSE, ANALYZE);\n\n```\n\n**STEP 4**.Again check the dead tubles\n\n\n\n```\npostgres=# select n\\_dead\\_tup ,last\\_vacuum,last\\_analyze,n\\_tup\\_upd, n\\_tup\\_del,n\\_tup\\_hot\\_upd,relname ,seq\\_scan,idx\\_scan from pg\\_stat\\_all\\_tables where relname='k1';\n n\\_dead\\_tup | last\\_vacuum | last\\_analyze | n\\_tup\\_upd | n\\_tup\\_del | n\\_tup\\_hot\\_upd | relname | seq\\_scan | idx\\_scan \n------------+-------------------------------+-------------------------------+-----------+-----------+---------------+---------+----------+----------\n 0 | 2017-05-15 13:48:05.985591-04 | 2017-05-15 13:48:16.555511-04 | 4708352 | 8192 | 8333 | k1 | 21 | \n(1 row)\n```\n\n**For more details: [Vacuum and analyze](https://www.tutorialdba.com/2017/05/definition-vacuum-reclaims-storage.html)**\n\n\n",
"name": "",
"is_accepted": true
}
] |
|
7c6e53ed2924b6795eea4400cd3f3e48 | How to restrict to connect the postgres user for sunday only ? | tutorialdba.com | 2020.16 | [
{
"text": "\nAlter a role so it does not allow login access on Sundays:\n\n\n\n```\nALTER ROLE user3 DENY DAY 'Sunday';\n```\n\n",
"name": "",
"is_accepted": false
}
] |
|
e429473157cfba252488dbbc1b007294 | One doubt I have... while we execute pg\_dump in backend there will be number of sessions will be created? | while we execute pg\_dump in backend there will be number of sessions will be created? | tutorialdba.com | 2020.16 | [
{
"text": "Check it crontab time if you set accidentally every one hours or any minutes ,for example if you set every one hour means if that dump size is large it will start again with other PID once one hour reached so dump will be overwritten, before scheduling crontab set it want to run one month once, other option if once job is run command that crontab job, other option run that pg\\_dump in background using & and nohub \n\n \n\nThen check the idle connection whether its comming over or not",
"name": "",
"is_accepted": false
},
{
"text": "And check the backup script any loop is there or not if any loop is there ,possiblity will be there automatically session creation",
"name": "",
"is_accepted": false
}
] |
ded908b04f18f2ba389e26838f691e5b | Enable PostgreSQL archivelog with different types | tutorialdba.com | 2020.16 | [
{
"text": "\nIn this tutorial i will explained about how to enable archivelog in postgresql server\n\n\n* WAL Archive log In PostgreSQL database system, the actual database 'writes' to an addition file called write-ahead log (WAL) to disk.\n* It contains a record of writes that made in the database system. In the case of Crash, database can be repaired/recovered from these records.\n* Normally, the write-ahead log logs at regular intervals (called Checkpoints) matched against the database and then deleted because it no longer is required. You can also use the WAL as a backup because,there is a record of all writes made to the database.\n\n\n**WAL Archiving Concept :**\n\n\n \n\n\n* In pg\\_xlog/pg\\_wal write ahead logs are stored. It is the log file, where all the logs are stored of committed and uncommitted transaction. It contains max 6 logs, and last one overwrites. If archiver is on, it moves there.\n* The write-ahead log is composed of each 16 MB large, which are called segments.\n* The WALs reside under pg\\_xlog/pg\\_wal directory and it is the subdirectory of 'data directory'. The filenames will have numerical(0-9) and character(a-z) named in ascending order by PostgreSQL Instance. To perform a backup on the basis of WAL, one needs a basic backup that is, a complete backup of the data directory, and the WAL Segments between the base backup and the current date.\n\n\n**Type 1 :**\n\n\n--First check the postgresql server whether archive log already enable or not using psql terminal.\n\n\n\n```\npostgres=# show archive\\_mode;\n archive\\_mode \n--------------\n off\n(1 row)\n\npostgres=# show archive\\_command;;\n archive\\_command \n-----------------\n (disabled)\n(1 row)\n```\n\n--Let’s create directories for backups and archives (login as root USER):\n\n\n\n```\nmkdir -p /Archive/Location\nmkdir /backups\nchown postgres:postgres -R Archive \nchown postgres:postgres /backups\n```\n\n-- check the folder permission\n\n\n\n```\n[root@p1 /]# ls -lrt\ndrwxr-xr-x 2 postgres postgres 4096 Apr 10 07:54 backups\ndrwxr-xr-x 3 postgres postgres 4096 Apr 10 07:54 Archive\n```\n\n-- switch to postgres user and modify postgresql.conf. Making the cluster into Archive-log mode requires RESTART.\n\n\n\n```\n[root@p1 /]# su postgres\ncd $PGDATA\necho 'max\\_wal\\_senders=1' >> postgresql.conf\necho 'wal\\_level=hot\\_standby' >> postgresql.conf\necho 'archive\\_mode=on' >> postgresql.conf\necho \"archive\\_command='cp %p /Archive/Location/%f'\" >> postgresql.conf\n```\n\n-- Restart the server and check the server whether running or not\n\n\n\n```\n**bash-3.2$ pg\\_ctl restart -D $PGDATA** \nwaiting for server to shut down.... done\nserver stopped\nserver starting\nbash-3.2$ 2017-04-10 08:12:07 EDT LOG: redirecting log output to logging collector process\n2017-04-10 08:12:07 EDT HINT: Future log output will appear in directory \"pg\\_log\".\n\n**bash-3.2$ pg\\_ctl status -D $PGDATA**\npg\\_ctl: server is running (PID: 9825)\n/opt/PostgreSQL/9.3/bin/postgres \"-D\" \"/opt/PostgreSQL/9.3/data\"\n```\n\n-- After the server start check the archive whether properly enable or not using psql terminal.\n\n\n\n```\npostgres=# show archive\\_mode;\n archive\\_mode \n--------------\n on\n(1 row)\n\n\npostgres=# show archive\\_command;\n archive\\_command \n----------------------------\n cp %p /Archive/Location/%f\n(1 row)\n\n\npostgres=# show wal\\_level; \n wal\\_level \n-------------\n hot\\_standby\n(1 row)\n```\n\n**Type II :**\n\n\nIf you want gzip format archives means you can give GZIP instead of CP then restart the server .\n\n\n\n```\narchive\\_mode =on # enables archiving; off, on, or always\n # (change requires restart)\narchive\\_command = 'gzip < %p /archivelog/archive\\_9.6/%f'\n```\n\n**Note :** Archiver won't enable when wal\\_level is \"minimal\" , wal\\_level should be **hot\\_standby** or **replica** if above 9.6.\n\n\n \n\n\n**Type III :**\n\n\nShipping archive to remote location .\n\n\n\n```\nwal\\_level = replica # minimal, replica, or logical\narchive\\_mode = on \narchive\\_command = 'rsync -av %p /home/postgres/archive/%f && rsync -av %p [email protected]:/home/postgres/archive/%f'\n```\n\n**Read More :** <https://www.tutorialdba.com/2018/06/how-to-setupconfigure-postgresql.html>\n\n\n",
"name": "",
"is_accepted": false
}
] |
|
0cd0b5daf0853a6329ed02cb95d0b8b3 | How to start and stop the postgres server automatically ? | tutorialdba.com | 2020.40 | [
{
"text": "\nbelow instructions for postgres auto stop and start. Please test first in testing server. Kindly let us know incase of issues during implementation. \n\nScript provided for postgres auto stop and start, but in below script we mentioned that script will be successfully tested in postgres server.\n\n\n \n\n1) Create scripts (postgres\\_start.sh,postgres\\_start.sh) in the path /etc/init.d/\n\n\n\n```\n====postgresql\\_start.sh=====\n#!/bin/bash\nrunuser -l insta -c '$PGHOME/bin/pg\\_ctl -D $PGDATA start > $PGDATA/start\\_log.txt.`date -I`' \n```\n\n\n```\n====postgresql\\_stop.sh====\n#!/bin/bash\nrunuser -l insta -c '$PGHOME/bin/pg\\_ctl -D $PGDATA stop > $PGDATA/stop\\_log.txt.`date -I`'\n```\n\n2) create a soft link in the runlevel to pick the script as a servcie.\n\n\n\n```\n<<Hostname>>:/etc/init.d/rc3.d # ln -s ../postgresql\\_stop.sh K60postgres\\_stop ===> for stop/kill\n<<Hostname>>:/etc/init.d/rc3.d # ln -s ../postgresql\\_start.sh S60postgres\\_stop ===> for start\n```\n\n==>>> Test it by rebooting the server \n\n\n \n\n \n\n\n",
"name": "",
"is_accepted": false
}
] |
|
c465286d6dd69c4d8fadd3137f8666c3 | In postgresql Can we append the data in schema from dump ? | tutorialdba.com | 2020.16 | [
{
"text": "\n--Yes you can restore the dump on specific schemap\n\n\n\n```\npg\\_restore --schema schema\\_name DBA.dump\n```\n\n--If you want to restore specific table into Postgresql schema\n\n\n\n```\npg\\_restore --schema schema\\_name -t table\\_name DBA.dump\n```\n\n",
"name": "",
"is_accepted": false
}
] |
|
513659e3d75ee0836392797c0a0045ff | как можно залогировать sql запрос, который строит nhibernate? | логирование sql запросов | barsup.net | 2020.45 | [
{
"text": "в appsettings.config / appsettings.local.config у ключей начинающихся с nhibernate, выставить уровень debug.\n\n“minimumlevel”: {\n\n“default”: “information”,\n\n“override”: {\n\n“nhibernate.cfg”: “debug”,\n\n“nhibernate.driver”: “debug”,\n\n“nhibernate.hql”: “debug”,\n\n“nhibernate.hql.ast.antlr”: “debug”,\n\n“nhi…",
"name": "",
"is_accepted": true
}
] |
30ab1c39104f160a30e4863c12bcd9dd |
var hl\_oldOnLoad = window.onload;
window.onload = function(){
if (typeof hl\_oldOnLoad == 'function')
hl\_oldOnLoad();
SyntaxHighlighter.autoloader(
'applescript ../qa-plugin/syntax-highlighter/scripts/shBrushAppleScript.js',
'actionscript3 as3 ../qa-plugin/syntax-highlighter/scripts/shBrushAS3.js',
'bash shell ../qa-plugin/syntax-highlighter/scripts/shBrushBash.js',
'coldfusion cf ../qa-plugin/syntax-highlighter/scripts/shBrushColdFusion.js',
'cpp c ../qa-plugin/syntax-highlighter/scripts/shBrushCpp.js',
'c# c-sharp csharp ../qa-plugin/syntax-highlighter/scripts/shBrushCSharp.js',
'css ../qa-plugin/syntax-highlighter/scripts/shBrushCss.js',
'delphi pascal ../qa-plugin/syntax-highlighter/scripts/shBrushDelphi.js',
'diff patch pas ../qa-plugin/syntax-highlighter/scripts/shBrushDiff.js',
'erl erlang ../qa-plugin/syntax-highlighter/scripts/shBrushErlang.js',
'groovy ../qa-plugin/syntax-highlighter/scripts/shBrushGroovy.js',
'java ../qa-plugin/syntax-highlighter/scripts/shBrushJava.js',
'jfx javafx ../qa-plugin/syntax-highlighter/scripts/shBrushJavaFX.js',
'js jscript javascript ../qa-plugin/syntax-highlighter/scripts/shBrushJScript.js',
'perl pl ../qa-plugin/syntax-highlighter/scripts/shBrushPerl.js',
'php ../qa-plugin/syntax-highlighter/scripts/shBrushPhp.js',
'text plain ../qa-plugin/syntax-highlighter/scripts/shBrushPlain.js',
'py python ../qa-plugin/syntax-highlighter/scripts/shBrushPython.js',
'ruby rails ror rb ../qa-plugin/syntax-highlighter/scripts/shBrushRuby.js',
'sass scss ../qa-plugin/syntax-highlighter/scripts/shBrushSass.js',
'scala ../qa-plugin/syntax-highlighter/scripts/shBrushScala.js',
'sql ../qa-plugin/syntax-highlighter/scripts/shBrushSql.js',
'vb vbnet ../qa-plugin/syntax-highlighter/scripts/shBrushVb.js',
'xml xhtml xslt html ../qa-plugin/syntax-highlighter/scripts/shBrushXml.js'
);
SyntaxHighlighter.config.stripBrs = true;
SyntaxHighlighter.all()
}
| How do I calculate sample size using a power | psychstat.org | 2020.40 | [
{
"text": "I am sorry we haven't compiled a complete manual to use our calculators. \n\n \n\nTo calculate the sample size using any of the calculators, you can input the desired power in the \"Power\" field and then remove the sample size in the \"Sample size\" field (leave it blank). Then the output will display the needed sample size according to the power you specified.",
"name": "",
"is_accepted": false
}
] |
4aa2b611a735766b4db597299fa2cd7e | Друзья, интересно было бы послушать ваше мнение на этот счет. Кто что думает и нравятся ли подобные арты, которые появляются в нашей всеобщей мировой паутине.
Немного о хентай: С японского этого слово переводится как «извращение». Это манга (комиксы) и аниме (японская анимация), в которых содержатся откровенные сцены сексуального характера. Само слово «хентай» используется преимущественно на Западе и в странах СНГ, японцы используют для обозначения комиксов и мультфильмов 18+ другие термины. Хентай – это:
Героини с милыми, кукольными, почти детскими лицами
Демонстрация фетишей, сексуальных извращений
Жанр японского искусства, получивший широкое распространение во всем мире
Источник: https://intrigue.dating/samoe-pikantnoe/payzuri-hentay-i-drugie-jelementy-yaponskogo-iskusstva-dlya-vzroslyh/© intrigue.dating
| Как вы относитесь к ветингу изображенному в стиле хентай? | omorashiclub.com | 2021.39 | [
{
"text": "Хорошая подборка. Арты чудненькие.\n ",
"name": "",
"is_accepted": false
},
{
"text": "\"\n \n\n\n Хентай – это:\n \n\n\n Героини с милыми, кукольными, почти детскими лицами\n \n\n \"\n \n\n\n Именно поэтому отношусь нейтрально. \n \n\n\n Предпочитаю видеть полностью сформировавшуюся девушку/женщину в этом пикантном положении, которое является темой форума.\n \n\n\n Кстати и \"обычное\" половое влечение возникает к фигуристым девушкам.\n ",
"name": "",
"is_accepted": false
},
{
"text": "Можно добавлять и свои арты в эту ветку обсуждений Эдуард. Было бы приятно посмотреть и те, которые нравятся именно вам\n ",
"name": "",
"is_accepted": false
}
] |
ef28c9826deaeb2b53bc9728c869513f | hello!
i am working on a project where i am gathering income components to create a social welfare measure. i am using the asec samples as my data set, including 2019. in gathering data for my project, i noticed that beginning with the 2019 asec, the cps disaggregates the retirement income measures… | incretir in asec 2019: why is the universe 58+? | ipums.org | 2019.51 | [
{
"text": "in 2019 the census bureau changed the universe of the incretir source variable, for unexplained reasons. however, they make the data available for people under age 58 as well. the variables [incret1](https://cps.ipums.org/cps-action/variables/increti1#description_section) and [incret2](https://cps.ipums.org/cps-action/variables/increti2#description_section) contain data from all individuals age 15+. summing these will give you a variable that is …",
"name": "",
"is_accepted": true
}
] |
4681fe685c67fe212a84dbb9f97ab772 | can you provide an easy-to-understand definition for the columns orig\_are, shape\_leng, and shape\_area on the gis files for the 1930-2000 tract (tiger 2008).
can you also provide a definition for area\_uncli?
the problem that i am having is that for each state - county i get multiple tract numbers (… | definitions for orig\_area, shape\_leng and shape\_area on 1930-2000 tract tiger 2008 gis files | ipums.org | 2021.17 | [
{
"text": "dear ivan,\n\nthanks for the added details!\n\nwe don’t include the centroid for census tracts in the shapefiles. you can compute the centroids using arcgis desktop (using the addxy function or something like that) or postgis (using the st\\_centroid function).\n\nbased on your description of your research …",
"name": "",
"is_accepted": true
}
] |
bad37db7bf2655d292433031a06cb663 | good evening,
why are there 53959 and 1447 cases in 1980 and 1990 census data that have 0 for wkswork2 and yet nonzero and nonmissing value on at least one of the following labor income components (incwage, incbus, incfarm)? are they simply data error, or is there any hidden information that i shou… | wkswork2=0 but have nonzero labor income | ipums.org | 2020.34 | [
{
"text": "that is close, but not quite what i meant.\n\nas you noted, there are no people with 0 values of wkswork2 who report income from these sources in 1970. the unexpected non-zero values you highlight in your initial post are much higher in 1980 than 1990. i cannot find anything in the underlying data or …",
"name": "",
"is_accepted": true
}
] |
e71e393b9965a0ea1c6c88e44eebf5a6 | ipums usa makes available the 1940 census 100% sample, where each row in the table represents a person. just to be specific, when i aggregate the person-level records for the state of georgia, i get 3,128,132 persons. however, various published summaries of the 1940 census put the population total i… | what explains the discrepancy between 1940 census ipums data and published summaries? | ipums.org | 2020.10 | [
{
"text": "you’ve guessed the correct answer. the original summaries (available through nhgis as you mentioned) were tabulated immediately following the enumeration, whereas the ipums usa microdata files are based on transcriptions of the microfilmed original enumeration forms decades later. while it may be sa…",
"name": "",
"is_accepted": true
}
] |
54d52c467cb383c8b2f3d6f23db5dde1 | do you have any suggestions for how to use yrimmig variable to determine exact year of immigration?
are there any imputation techniques recommended by ipums? the codebook notes comparability issues across survey years and suggests using last year in interval as conservative estimate. | exact year of migration and yrimmig variable | ipums.org | 2021.21 | [
{
"text": "there is no way to determine the exact year of immigration in the public use cps data. as you note, the comparability documentation suggests using the provided code as it is the latest possible year for each range as a conservative estimate. you may wish to compare your results using the latest poss…",
"name": "",
"is_accepted": true
}
] |
2d08aebcd4e922724e54d79af0c8079c | hi,
i am trying to create an extract with all the original variables (and all the observations) from the 1991 census of brazil. is there a way to select this automatically? if not, is there a way to at least list all the original variables from that census, so that i can use as a guide when adding … | extract brazil 1991 sample with all the original variables | ipums.org | 2020.05 | [
{
"text": "there is no way to “automatically” select all of the original source variables for any given census sample. the best way to go about this task is to select the source variables radio button on the [select data page](https://international.ipums.org/international-action/variables/group?variabletype=source) and then either go through all of the variable categories in the dropdown menus or sel…",
"name": "",
"is_accepted": true
}
] |
122e205de9c5d70c7e1785b6b3f9bb4b | spmnumchild is [defined in ipums as](https://cps.ipums.org/cps-action/variables/spmnchild#description_section) “the number of children in the family unit.”
it corresponds to spm\_numkids in the raw asec, which also [defines it](https://www2.census.gov/programs-surveys/cps/techdocs/cpsmar19.pdf) as “spm unit’s number of children.”
as [spm reports](https://www.census.gov/content/dam/census/library/publications/2019/demo/p60-268.pdf) count the child poverty rate as the poverty rate for people under age 18, i assumed that would be … | how is spmnumchild defined? | ipums.org | 2019.51 | [
{
"text": "hello max,\n\nyou’re picking up a very subtle definition. here’s the definition:\n\n“the term “children,” as used in tables on living arrangements of children under 18, are all persons under 18 years, excluding people who maintain households, families, or subfamilies as a reference person or spouse.” ([h…](https://www.census.gov/programs-surveys/cps/technical-documentation/subject-definitions.html#children)",
"name": "",
"is_accepted": true
}
] |
0512797c6f39591628b79ddccb7ac388 | i want a detailed extract with each activity as an individual variable - i’m trying to match the bls survey pdf (eg. <https://www.bls.gov/tus/tables/a1_2003.pdf>) structure which shows a full detail/heirarchy of time use activities.
eg. i want to see the time for ‘general or other personal care’ sepa… | how can i extract all subcategories in detail? | ipums.org | 2020.16 | [
{
"text": "you can add all the bls categories of time use by checking all the variables in the group on [this page](https://www.atusdata.org/atus-action/time_use_variables/group/3).",
"name": "",
"is_accepted": true
}
] |
0f8a1ccb2a1856ed5612384590740c56 | hi! is there an estimated timeline available for when the 2010 to 2020 geographic crosswalks will be available for helping compare the old and new census geographies? thanks a ton for all you do! | timeline for release of 2010 to 2020 geographic crosswalks for new census boundaries | ipums.org | 2021.25 | [
{
"text": "thanks for your interest!\n\nour next planned release will provide a crosswalk from 2020 blocks back to 2010 blocks, and we aim to publish that at least a week in advance of august 16, when the census bureau [plans to release](https://2020census.gov/en/important-dates.html) the first version of the 2020 redistricting data. if time allows, we may incl…",
"name": "",
"is_accepted": true
}
] |
04551a7054d7353915afeaed47d5c542 | is there a consensus on whether it is appropriate to combine census samples for the sample year? given that the jurisdictions and variables of interest are the same, can i combine, say a 1% sample and a 5% sample to yield a six percent sample? someone asked a similar question here: [questions about 1…](https://forum.ipums.org/t/questions-about-1970-census-data/3133) | combining census samples for a given year | ipums.org | 2021.04 | [
{
"text": "i would not assume that this is going to be the case for all samples. each census year, the public use microdata samples were selected using different methodologies. the details can be found at the links [here](https://usa.ipums.org/usa/voliii/tsampling.shtml).\n\nin cases where two samples were drawn independently, as in the example you gave, you’d ge…",
"name": "",
"is_accepted": true
}
] |
cee1cbfc9501e5ddc0c8b26c4000bb49 | i am trying to download an extract of the acs 2014-18 5-year sample, restricted to california and ages 25-54 in order to save space. the dialog box for “select cases” doesn’t include state fips code (st) or age (agep) as options. in fact, i only see 6 variables out of about 60: gq, race, raced, hisp… | can't "select cases" for most variables in extract of acs 2014-18 5-year sample, even though they are available in all years | ipums.org | 2020.40 | [
{
"text": "i see a large number of source variables in your latest extract request. you can only select cases for harmonized variables that you have included in your data cart. if you add the harmonized variables [statefip](https://usa.ipums.org/usa-action/variables/statefip#description_section) and [age](https://usa.ipums.org/usa-action/variables/age#description_section) in your cart (instead of the source variables), you should be able to restrict yo…",
"name": "",
"is_accepted": true
}
] |
e29fcd0b74d29db0798b2c6b2beeeda1 | i downloaded census tract level data and need to link them to the msa level for a few indicators. is there a way i can extract data in this format or this has to be done manually? i would also need to do the same for the shapefiles as i would run some gis analysis. any help is very appreciated.
tha… | linking census tract data to msa | ipums.org | 2021.39 | [
{
"text": "thanks for sharing the example. i had previously checked a 2020 census file (from nhgis dataset 2020\\_pl94171), and in that case, the cbsaa field contains the correct codes. i’ve now checked tract-level files from several other datasets, and it appears that, generally, tract-level files from decennia…",
"name": "",
"is_accepted": true
}
] |
a31117a1853f4db36d7fcd1631c0cb7e | i’m trying to obtain unemployment data (5-yr estimates 2014-2018) by census tracts in california for all race/ethnicity groups that are mutually exclusive from hispanics (i.e. non-hispanic whites, non-hispanic blacks, non-hispanic asians), however, all the tables that i’ve seen that come up in my se… | unemployment data by census tracts in ca for all race/eth groups that are mutually exclusive from hispanics | ipums.org | 2020.50 | [
{
"text": "i do not believe employment status broken down by mutually exclusive race and hispanic origin for the 2014-2018 5-year sample is available in nhgis (though non-exclusive groups are available, e.g. white only, black only, hispanic/latino). information on mutually exclusive hispanic groups by race is …",
"name": "",
"is_accepted": true
}
] |
f9f778cc3560ebd12d152a8a095221c6 | there seem to be msas missing from the [met2013 omission and commission errors by msa](https://usa.ipums.org/usa/resources/volii/met2013_match_errors_2010.xls) document. flint, mi, for example, is not included. however, it doesn’t appear to be a matter of not listing msas with zero omission or commission errors, either—flagstaff, az, for example, has zero such errors and … | msas missing from met2013 match errors file | ipums.org | 2020.10 | [
{
"text": "there are three met2013 omission and commission spreadsheets provided at the bottom of the [met2013 description](https://usa.ipums.org/usa-action/variables/met2013#description_section). they each report errors for a different puma vintage (2000 or 2010) and/or a different set of samples. in each case, they list only the msas that are identified by met2013, i.e., those msa…",
"name": "",
"is_accepted": true
}
] |
2c67be80e6460259b47e7321f11e1208 | i am currently working with asec linked files. in 1990 and 1991, i found a couple of incwage = 99999.
given the info available [here](https://cps.ipums.org/cps/topcodes_tables.shtml), i thought the topcode in those years was 199,998.
is this a typo? | incwage: error in posted 1988-1995 topcode? | ipums.org | 2021.39 | [
{
"text": "you are correct that the topcode for incwage in 1990 and 19991 is incwage = 199,998. incwage values are a combination of [inclongj](https://cps.ipums.org/cps-action/variables/inclongj) and [oincwage](https://cps.ipums.org/cps-action/variables/oincwage) (see the [description tab for incwage](https://cps.ipums.org/cps-action/variables/incwage#description_section) for more information), the topcodes of which for both is 99,999. this means that in cases where both inclongj and oincwa…",
"name": "",
"is_accepted": true
}
] |
e69a584df16194eb18f17a68e0b5f670 | hello,
i am trying to replicate the number of families in poverty from cps 2018 published in table 4 here: <https://www.census.gov/data/tables/time-series/demo/income-poverty/historical-poverty-people.html>. i am not able to replicate the number of families (with our without children) in the universe… | replicating family poverty estimates | ipums.org | 2019.47 | [
{
"text": "i think the issue is with your sample. the 2018 published numbers come from the 2019 asec. this is because most asec questions ask about the previous calendar year. re-run the numbers with 2019 asec and see if it replicates the census publication.",
"name": "",
"is_accepted": true
}
] |
63788e83c67e892121b5c24922d1870b | hello. i am trying to link basic monthly cps data across four consecutive months and check how good my matches are by looking at cpsidp linkages with age, sex, and race using stata. my ultimate goal is to calculate the quarterly transition rate of individuals from one industry to another industry fr… | linking basic months data across four consecutive months | ipums.org | 2021.10 | [
{
"text": "your general understanding is correct. there is a [whole page](https://cps.ipums.org/cps/cps_linking_documentation.shtml) on linking the cps, including many resources. there are also many materials from a summer 2018 workshop on using the cps longitudinally, [here](https://cps.ipums.org/cps/cps_workshop2018_materials.shtml). that page includes sample code for linking and validation of links. the way you’ve described you…",
"name": "",
"is_accepted": true
}
] |
2f3641405bb93b9587ac4903ac923095 | hello there,
it is mrinal from india looking for an easy way to get the merge data of all four nfhs. however, i found out that the very basic variable district does not have data on all four surveys.
why is the district data of 2005 nfhs not available on ipums?
is it desirable to have that piece … | district data of third nfhs (dsh, india 2005)? | ipums.org | 2019.43 | [
{
"text": "this is a good question. as is noted on the description tab for the [districtia](https://www.idhsdata.org/idhs-action/variables/districtia#description_section) variable: “ipums-dhs users should note that the india 2005 sample does not contain district level information due to confidentiality requirements of hiv testing.” therefore, the identification of district of residence for…",
"name": "",
"is_accepted": true
}
] |
a2480dd3f5d7d5a646ec41cee2fdb2ab | hello,
i’m currently attempting to match bls estimates for overall union membership between 2000 and 2020 using ipums cps microdata. i’ve been able to match the bls estimates spot on for 2003 to 2020.
however, i’m running into an issue with the estimates pre-2003. when i produce the estimates for … | matching bls union membership estimates | ipums.org | 2021.39 | [
{
"text": "you are indeed correct that there is a <1% difference between bls and ipums estimates for union membership in the 2000, 2001, and 2002 cps. this discrepancy has to do with cps weights that were re-calculated after the 2000 census (the weights that are currently available through ipums cps). unfortun…",
"name": "",
"is_accepted": true
}
] |
5e4016acf417109997c77eb705590133 | hi, do column locations in the extract codebooks refer to the locations of the variables in the extract or in the original raw data? that is, if i ever needed to reproduce the extract data “by hand” from the raw atus data, could i do it using the column locations provided in the extract codebooks? t… | column locations in atus-x extracts | ipums.org | 2020.10 | [
{
"text": "in the codebook file, provided with ipums extracts, the “column” indicate variable locations in the fixed-width data file extracted from ipums and not the original raw data. this information is useful when converting the fixed-width data file provided by ipums into a usable file format within your p…",
"name": "",
"is_accepted": true
}
] |
7551dfa3ff3e2e6c85db6a0fa664f3b8 | dear ipums,
i checked the proportion of observations whose incwage is topcoded in the 1940 full count data (below).
. fre incwage
incwage -- wage and salary income
-----------------------------------------------------------
| freq. percent valid cum.
-------------… | 1940 full count top code incwage | ipums.org | 2021.39 | [
{
"text": "regarding manual recoding of individuals to niu: ipums usa doesn’t generally enforce universe rules if they are not perfectly followed by the original microdata. because of this, assigning niu codes based on the stated universe “persons age 14+, not institutional inmates” is probably not going to ex…",
"name": "",
"is_accepted": true
}
] |
8ca45205d7c708d791ef5f0f38de09f1 | the ipums cps variable description for faminc notes:
faminc reports annual family income, in categories, of all persons related to the head of household/householder. for individuals who are not part of the householder’s family, faminc reports the value for the householder’s family. this measure in… | faminc includes income of all household members? | ipums.org | 2020.16 | [
{
"text": "you are correct about [faminc](https://cps.ipums.org/cps-action/variables/faminc#description_section) being a specific response on the cps questionnaire and [hhincome](https://cps.ipums.org/cps-action/variables/hhincome#description_section) being the sum of [inctot](https://cps.ipums.org/cps-action/variables/inctot#description_section) for all household members. the note about hhincome being the sum of all “adult” household members comes from the [universe for inctot](https://cps.ipums.org/cps-action/variables/inctot#universe_section), which (since 1980) includes only those age 15 and…",
"name": "",
"is_accepted": true
}
] |
fce01c9017113a83fae9b74378b41c4c | hello,
i am unsure about how to build an annual estimate from the cps basic sample.
for example, the bls reports: [(unadj) civilian labor force level - native born, with own children under 18, men](https://beta.bls.gov/dataviewer/view/timeseries/lnu01073449). i know how to compute monthly estimates for this measure. however, i am unsure about how to compute a… | annual estimate from basic sample | ipums.org | 2021.39 | [
{
"text": "generally to calculate an annual estimate, you would keep all records (including households appearing multiple times with different values of mish), use wtfinl, and divide the weights in each month by 12. each sample (month) is designed to be representative of the population (after weights are appli…",
"name": "",
"is_accepted": true
}
] |
eaa8986d32d21fa7428945ff6f840117 | hello,
i was curious whether the unharmonized bms labor-layoff variables, such as uh\_layavl\_b1 or uh\_laydt\_b1, would be released for november 2020 or not. if they are just delayed coming out sorry for being impatient!
-joel | layoff unharmonized variables | ipums.org | 2021.04 | [
{
"text": "these variables should now be available; we have corrected a technical error that caused them to display as unavailable in the extract system.",
"name": "",
"is_accepted": true
}
] |
d10a2bb0d7bfd73c7b617b2068a5d244 | i am trying to replicate the published estimate of the proportion of people living in poverty according to the one-year 2018 acs. we have been able to replicate the 13.1 percent proportion of people in poverty that was published here: [https://data.census.gov/cedsci/all?q=&g=&t=poverty&table=s1701&ti…](https://data.census.gov/cedsci/all?q=&g=&t=poverty&table=s1701&tid=acsst1y2018.s1701&hidepreview=false&lastdisplayedrow=16) | replicating 2018 acs poverty estimates | ipums.org | 2020.05 | [
{
"text": "this is likely due to the fact that the public use microdata files do not utilize the full acs sample. the tables on [data.census.gov](http://data.census.gov/) use the full acs sample, which is only accessible internally to census bureau researchers. the weights are not designed to sum to population totals for every possible …",
"name": "",
"is_accepted": true
}
] |
7db9037b1a3895ad1152c841ff0cf050 | i was computing month-to-month transitions between occupations (different digits as well as major and detailed groups) based on the harmonized occ2010 variable and noticed a weird spike in june 2015. it jumps to around 7% while between 2005 and today it is usually around 2% or 3%. also for the rest… | occ2010 any change in june/july 2015? | ipums.org | 2020.45 | [
{
"text": "the ipums cps team has looked into this issue and it appears to be an undocumented anomaly in the original public use microdata files obtained from the census bureau. thanks for pointing this out. unfortunately, we are unable to correct these issues in the underlying data, but we will update the doc…",
"name": "",
"is_accepted": true
}
] |
1320017a12bb54137e04638f66dbf54f | i have just got to learn about the cws and would like to know if the cws and asec could be merged. if that’s possible, could you let me know which id i need to use for data merge and which weight for the analyses? thanks. | linking contingent worker supplement to asec? | ipums.org | 2020.45 | [
{
"text": "you can indeed link about half of the people in the cws (which is in may) to the asec. to learn more about possible linkages in the cps (and the percentage of the sample you can expect to link across months), i would encourage you to visit the [cps rotation pattern explorer (ropes) tool](https://cps.ipums.org/rotation_pattern_explorer#/). the variable…",
"name": "",
"is_accepted": true
}
] |
f7e1e3f5f01f4856c8e2294ade71352b | the documentation indicates that cutoff and offcutoff should differ for some asec records, at least for secondary individuals under 15 who are in the [cutoff universe](https://cps.ipums.org/cps-action/variables/cutoff#universe_section) but not the [offcutoff universe](https://cps.ipums.org/cps-action/variables/offcutoff#universe_section).
however, these variables are identical in the 2020 asec, and 0.2% of records have the 999999 niu code… | cutoff and offcutoff are identical in 2020 asec | ipums.org | 2021.39 | [
{
"text": "the census bureau updated their poverty universes in 2020, removing differences between official poverty variables (i.e, [offcutoff](https://cps.ipums.org/cps-action/variables/offcutoff#description_section)) and non-official poverty variables (i.e. [cutoff](https://cps.ipums.org/cps-action/variables/cutoff#description_section)). as you can see in table 1 on [this page](https://cps.ipums.org/cps/poverty_notes.shtml), the official poverty rate and the fixed poverty rate are the same. thank you f…",
"name": "",
"is_accepted": true
}
] |
e57bc4a385992f80ce868a301776d7fe | hi, i think i understand the problem and the need for this variable, but i’m still a little unclear on what the mismatch is reporting in [met2013err](https://usa.ipums.org/usa-action/variables/met2013err#description_section) variable. you aggregate census block data within pumas thought to be located within a metro area to see how large of a mismatch there is in between the … | met2013err - what is the mismatch reported exactly? | ipums.org | 2021.31 | [
{
"text": "we get the populations for calculating mismatches from census summary data–not microdata. the census bureau provides files that identify the 2000 populations of the parts of 2000 pumas, and the 2010 populations of the parts of 2010 pumas, so we can use those. but they don’t provide any info about th…",
"name": "",
"is_accepted": true
}
] |
c55bb211d36a052bf3f99880cac8c76e | hi, i have data from atus respondents only, all years, and am trying to understand the niu (not in universe) coding.
shouldn’t there be niu observations for time use variables re: children for those with no children in the household, instead of 0 values for those who don’t have household kids? fo… | questions about niu observations | ipums.org | 2020.16 | [
{
"text": "i will aim to address each of your questions one at a time.\n\n(1) if your data only include respondents, then everyone in your data is eligible for all time use questions. one way to ensure that time spent with children excludes respondents with no children is to include the [kidund18](https://www.atusdata.org/atus-action/variables/kidund18#description_section) variable in your…",
"name": "",
"is_accepted": true
}
] |
3acac92fcfb318352847e72a9668b5eb | hi friends,
i was checking the data quality flag variable [qmhmar1](https://cps.ipums.org/cps-action/variables/qmhmar1#codes_section). qmhmar1 is a data quality flag for mhmar1m and mhmar1y (i.e., year and month of first marriage). this variable is part of the cps fertility supplement.
why are the number of cases with “[allocated](https://cps.ipums.org/cps-action/variables/qmhmar1#codes_section)” values (qmhmar1=4) so high in yea… | qmhmar1 in years 86-88 | ipums.org | 2021.39 | [
{
"text": "the ipums cps team has looked into the 1986 - 1988 frequencies and agrees that they are inconsistent with frequencies from surrounding samples. there is insufficient evidence, however, for them to determine whether or not there is a coding error in the original cps data, so these data are being left…",
"name": "",
"is_accepted": true
}
] |
e835d399f161784e078e337648f771a4 | let me preface this question with an acknowledgement that there has been a lot of conversation on this forum about how to decode the differences between top codes, replacements, and swap values. i have combed through the postings about this and i am still having a difficult time getting my head arou… | maximum possible value v. replacement / swap values | ipums.org | 2020.16 | [
{
"text": "i think a quick clarification will help clear up the confusion here. in the case when the swap value equals the maximum possible value (as is the case for [incsurv1](https://cps.ipums.org/cps-action/variables/incsurv1#codes_section) and [incsurv2](https://cps.ipums.org/cps-action/variables/incsurv2#codes_section) in 2016 and 2018), no values are swapped and anyone above the swap value threshold simply receives the maximum possible val…",
"name": "",
"is_accepted": true
}
] |
7c96f158fbc16b1bcb9e2e0475b8c337 | i have a question and two suggestions on the asec variables capgain and caploss:
question:
why have these variables been discontinued since 2008? (see their availability tabs.)
the reason i am asking is threefold:
a) all other tax variables imputed by the cb tax model are available for later yea… | capgain caploss: question and suggestions | ipums.org | 2021.17 | [
{
"text": "the best explanation i could find for why these variables were discontinued is in user note 3 from the [2009 asec documentation](https://www2.census.gov/programs-surveys/cps/techdocs/cpsmar09.pdf) (see the last page):\n\ndata for two “after-tax” variables, cap\\_gain and cap\\_loss are not included in the 2009 asec public use file. they were removed from the “after-tax” mo…",
"name": "",
"is_accepted": true
}
] |
e92e3f20f203dc2fe61df4335b08ede2 | hello,
i’m currently attempting to match the family status estimates in [table h-01](https://www.census.gov/data/tables/time-series/demo/income-poverty/cps-hi/hi.html) of the census’ tables on health insurance coverage and would love some insight into what i am potentially getting wrong!
specifically, i’m interested in the values in column b of the excel workbook (the “total” colu… | matching census family status estimates | ipums.org | 2021.25 | [
{
"text": "i believe you should be defining families using [ftype](https://cps.ipums.org/cps-action/variables/ftype#codes_section)=1,3 and unrelated individuals as ftype=2,5. ftype gives the family or subfamily type, while famrel gives the status of individuals within the family (useful for identifying householders). here’s what i calculated:\n\ntable h-01\n\nipums\n\n% diff\n\n…",
"name": "",
"is_accepted": true
}
] |
7d4e8afbfb43f0ebfb322d41218a0c49 | dear all,
i have a question regarding the census 2000 and 2010. i was looking at the samples:
|[2000 1%]|1.0%||
|[2010 acs 5yr]|5.0%||
|[2010 10%]|10.0%|
and specifically, i´m looking at the metro area “san diego-carlsbad, ca”. which is identified in the variable met2013 for the 2010 year. but … | met2013 and metarea variables in census 2000 and 2010 | ipums.org | 2021.10 | [
{
"text": "as it happens, for san diego specifically, metarea and met2013 are directly comparable because the san diego metro area (labeled as “san diego-carlsbad” in the 2013 definition) has always covered the same geographic extent throughout its history, and both metarea and met2013 are able to identify tha…",
"name": "",
"is_accepted": true
}
] |
675e3705c17e946f9fdd872f0a0a6d52 | reviewing the ipums [income topcode table](https://cps.ipums.org/cps/topcodes_tables.shtml), i have noticed that for the variables [incss](https://cps.ipums.org/cps-action/variables/incss) and [incssi](https://cps.ipums.org/cps-action/variables/incssi) there appears to be no registered topcode for the asec 2011 and beyond.
in the ‘all income components topcode by year’ section, both incss and incssi have entries up through the 2003-2010 column. howeve… | are incss and incssi topcoded values swapped in asec 2011-? | ipums.org | 2019.51 | [
{
"text": "this is a good question. looking through the asec codebooks for post-2010 samples, i do not see incssi or incss listed in the “rank proximity swap” tables. this implies that these variables are not topcoded in post-2010 asec samples. you can inspect the asec codebooks for each year on [this page](https://cps.ipums.org/cps/codebooks.shtml#asec_codebooks).",
"name": "",
"is_accepted": true
}
] |
a23457dd0333844bc11547af7760f719 | is there a good way to identify whether information on wages hourly wages or weekly earnings from the cps org was collected by self or proxy response?
i see that lfproxy contains information on whether labor force information was collected by self or proxy response.
could i use this variable to id… | identifying self-reported wages in the cps org | ipums.org | 2021.43 | [
{
"text": "[lfproxy](https://cps.ipums.org/cps-action/variables/lfproxy#description_section) should be appropriate for identifying who reports labor force information in the basic monthly survey for cps; i think this should also include the org questions. this is based on what i am seeing in the [questionnaires](https://www2.census.gov/programs-surveys/cps/techdocs/questionnaires/labor%20force.pdf), but want to be clear that i cannot find explicit documentation confirmin…",
"name": "",
"is_accepted": true
}
] |
aab5d2190aa4716dbff5c1a955ecb399 | hello - i’m interested in how economic conditions have changed over time from 2000 to 2015, so i’m using the geographic crosswalk files as specified with gisjoiner and the interpolation weights. the instructions are pretty straightforward when my variable of interest is a count (e.g., # of foreign b… | using medians instead of count w/geographic crosswalks | ipums.org | 2021.39 | [
{
"text": "using the nhgis crosswalks for medians is not as simple. (this is the main reason we haven’t yet extended nhgis [standardized time series](https://www.nhgis.org/documentation/time-series#standardization) to include medians.) e.g., if you applied the crosswalk weights to median household income in exactly the same way as with household counts, then if a target unit …",
"name": "",
"is_accepted": true
}
] |
eb6a905d0b75ec22f08da79edb3840d6 | with the 2019 acs 1yr pums data, i calculate 122,802,904 households with gq in (0,1,2,5) (same for gqtype in (0,9)), but [data.census.gov](http://data.census.gov) shows 122,802,852. can anyone shed light on this discrpancy of 52 households? | discrepency in 2019 us households | ipums.org | 2021.17 | [
{
"text": "we do not generally expect to exactly replicate official us census statistics with public use data, though we expect estimates to be within the margin of error around the official estimates. official statistics rely on more detailed versions of the data than the public use versions available through…",
"name": "",
"is_accepted": true
}
] |
7bbf19b4553de5138dfa4951efb6f47d | hi michelle, why are flag variables such as [qincwage](https://usa.ipums.org/usa-action/variables/qincwage) not available in 1940 full count data? thanks. | flags in 1940 full count data | ipums.org | 2021.39 | [
{
"text": "qincwage does not exist for the 1940 full count sample because there were no allocated or imputed values for incwage in that dataset. that is the case with most variables in the full count census files, unless they were specifically edited by the ipums usa team.",
"name": "",
"is_accepted": true
}
] |
bec2296bd6f8782ba0365ac3093b74e4 | as of the 2019 asec, two income components are introduced that measure income accrued from interest on retirement accounts: [incrint1](https://cps.ipums.org/cps-action/variables/incrint1#description_section) and [incrint2](https://cps.ipums.org/cps-action/variables/incrint2#description_section). how do these components relate to [incint](https://cps.ipums.org/cps-action/variables/incrint1#description_section)? is incint inclusive of the values from incrint1 and incrint2, or are they mutually exclusive? | incrint1 versus incint | ipums.org | 2020.05 | [
{
"text": "incint includes incrint1 and incrint2. you can see which source variables correspond to which ipums variables for each sample using the tables on [this page](https://cps.ipums.org/cps/topcodes_tables.shtml).",
"name": "",
"is_accepted": true
}
] |
8a30988c164b5e6d629e3eabcf9b1aac | i want to use the 2007 ethiopia census for analysis. i wonder, however, at which admin level the census provides representative information. the sample overview information states that the geographical unit is the zone, which is admin level 2. however, when i look at various census documents and at … | geographic unit ethiopia 2007 census | ipums.org | 2021.04 | [
{
"text": "[wereda](https://international.ipums.org/international-action/variables/werdet#codes_section) (admin level 3) is available in ipums for the ethiopia 2007 sample and provides representative information. it likely was not listed on the [sample characteristics page](https://international.ipums.org/international-action/sample_details/country/et#tab_et2007a) because ipums international does not spatially harmonize level 3 geography. if you use wereda for your analysis, keep in mind…",
"name": "",
"is_accepted": true
}
] |
f16c98b48d4d442faab44e2715034959 | i’m looking at marital status and hours of work of the spouse for asec samples and find some inconsistencies. for example, variable uhrsworkly\_sp (hours of work of the spouse last year) pernum 2, serial 41938 in years 2007 and 2008. (i understand that they are different individuals, i’m just using t… | spouse hours of work and marital status | ipums.org | 2020.40 | [
{
"text": "i agree with your assessment of the 2007 case where the person is married but their spouse has a uhrsworkly value of 999; the spouse may not have worked last year. the second situation is unmarried partners. their marital status (appropriately) reflects that they are not married (and specifically ha…",
"name": "",
"is_accepted": true
}
] |
67f519cbac55e1c528f9e8da16254d0f | i understand that certain of the income variables in the cps asec are entirely imputed, meaning that they are equal to an equivalent dollar value rather and that their values are not derived from a self report by respondent. for example, a response from an ipums staff member to [this post](https://forum.ipums.org/t/what-does-family-market-value-mean-w-r-t-schllunch/2255) helpfully c… | imputed variables with 'market value' | ipums.org | 2019.51 | [
{
"text": "you are correct that [ffngcare](https://cps.ipums.org/cps-action/variables/ffngcare#description_section) and [pmvcaid](https://cps.ipums.org/cps-action/variables/pmvcaid#description_section) refer to the market value for these variables. these values are “imputed” based on responses by individuals to the cps questionnaire. the [stimulus](https://cps.ipums.org/cps-action/variables/stimulus#description_section) variable was a new variable included only in the 2009 asec sample and is a product of the cps tax model. the sa…",
"name": "",
"is_accepted": true
}
] |
9f4cf3697a15bbe6f48390c162bb9570 | hi!
i am using mexico 2010 data and doing an individual-level analysis.
i read all the available files about mexico 2010 survey design, and i have a few doubts. i am using stata to analyze the data, and therefore i need to declare the survey design.
i found that it was a one-stage stratified clu… | sampling design mexico 2010 | ipums.org | 2020.34 | [
{
"text": "this is a one-stage stratified sample survey, stratified by municipality and clustered by enumeration area ([link](https://international.ipums.org/international-action/sample_details/country/mx#tab_mx2010a) for this sample’s design info). the psu is the enumeration area, and it’s currently only available as a source variable in ipums-i ([mx2010a\\_pmu](https://international.ipums.org/international-action/variables/mx2010a_0071) and [mx2010a\\_pmup](https://international.ipums.org/international-action/variables/mx2010a_0483), which are identical). the…",
"name": "",
"is_accepted": true
}
] |
c158db8be7ce931e12d068e4b6eac2d0 | hi, i’m using atus data from all available years. are the earnings variables in the atus extracts, e.g. ‘earnweek’, adjusted for inflation? if not, are the corresponding cps variables adjusted for inflation? thanks much, sanjiv | atus earnings adjusted for inflation? | ipums.org | 2020.16 | [
{
"text": "as noted on the [comparability tab](https://www.atusdata.org/atus-action/variables/earnweek#comparability_section) of the earnweek variable, this variable is not adjusted for inflation; neither are the corresponding ipums cps variables. ipums cps does, however, provide a [user note](https://cps.ipums.org/cps/cpi99.shtml) on adjusting for inflation.",
"name": "",
"is_accepted": true
}
] |
d3deb9e7ae7adc3f5ae0d2c17035d5cc | i am wondering if it is appropriate to calculate state estimates with the linked atus-cps data. for example, i want to look at the leave use module outcomes from the atus in a single state. i have weighted the data with the lvwt weight and 160 replicate weights to get national level estimates but am… | atus-cps linked state estimates | ipums.org | 2020.29 | [
{
"text": "we asked our contacts at bls and they recommend pooling 5 years of atus data to generate state-level estimates. there aren’t 5 years of leave module data, so state-level estimates for that module are not recommended.",
"name": "",
"is_accepted": true
}
] |
65dfc8c1f96e793f6181ece423044c56 | the [faq on acs/prcs multi-year files](https://usa.ipums.org/usa/acs_multyr.shtml) states
for some high-income cases, previously topocoded values are no longer topcoded.
does this only apply to income variables? (is there a comprehensive list of affected variables?)
(i compared property taxes in the 2009 5 yr acs file to the single year fi… | topcodes in multiyear file | ipums.org | 2021.25 | [
{
"text": "for ipums multi-year datasets, variables reporting dollar values have been inflated to dollars for the last data year resulting in top- and bottom- codes that differ from the single-year acs documentation. this difference between the ipums 1-year files and ipums 5-year files applies to person-level …",
"name": "",
"is_accepted": true
}
] |
abfe9b5469b6ddd42b593823cb28c52f | hi, i am looking for homeless count data from the 2000 decennial census. i was not able to locate these counts using the nhgis data explorer even though published papers state the data is available from nhgis at the census tract level. i have examined all relevant tables (such as group quarters popu… | 2000 dec census homeless count data at cen tract level | ipums.org | 2020.45 | [
{
"text": "after some investigative work, i discovered that the census bureau did not release homeless count data for the 2000 dec census (see holmes 2001 [nyt article](https://www.nytimes.com/2001/06/28/us/bureau-won-t-distribute-census-data-on-homeless.html)). rather the census bureau released a count of “other non-institutional group quarters” in the pct 16 and 17 data tables (sf-1). i was referenci…",
"name": "",
"is_accepted": true
}
] |
07d082c658e852e79372d9242282517c | in the sample selection page i can choose asec samples from 1962 on. however, when i examine the data i downloaded, i can seen that asecflag is only defined for years 1976 onwards. i couldn’t find much information in the [documentation](https://cps.ipums.org/cps-action/variables/asecflag#description_section) provided for that variable. why is asecflag defined only for 1976… | cps asecflag before 1976 | ipums.org | 2020.34 | [
{
"text": "asecflag indicates whether the respondent is part of the asec or march basic sample, according to the [documentation](https://cps.ipums.org/cps-action/variables/asecflag#description_section). if you look at the [availability of samples](https://cps.ipums.org/cps-action/samples), asec is available from 1962 onwards (under the asec tab), whereas basic monthly is available from 1976 onwards (under the basic monthly tab…",
"name": "",
"is_accepted": true
}
] |
c06129f954b259c65dc6b1ce1a2477ae | hi all,
i’m trying to find out why i can create an extract of ipums usa 2019 acs data that includes the variable “pwstate2” but i cannot seem to locate this variable in the online data analysis too (abacus).
does abacus not provide all variables otherwise available to extracts?
when i use it and … | missing variables in abacus (online analysis) | ipums.org | 2021.17 | [
{
"text": "abacus only includes a subset of the variables available through ipums usa. however, pwstate2 is available via [sda, another online analysis tool](https://usa.ipums.org/usa/sda/) available for analyzing ipums data (you will need to scroll down a bit further on the page i linked previously to access the sda information). i am also li…",
"name": "",
"is_accepted": true
}
] |
1af1249f1407b3d78722c4ae86b6c8ba | are the maximum possible values for stampval the same as for foodstamp? i haven’t been able to find information on this in neither the [stampval description](https://cps.ipums.org/cps-action/variables/stampval#description_section) nor the [topcoding tables](https://cps.ipums.org/cps/topcodes_tables.shtml).
thanks for clarifying! | stampval maximum possible value | ipums.org | 2020.40 | [
{
"text": "thank you for clarifying and my apologies for the misunderstanding.\n\nthe comparability tabs for [foodstamp](https://cps.ipums.org/cps-action/variables/foodstamp#comparability_section) and [stampval](https://cps.ipums.org/cps-action/variables/stampval#comparability_section) indicate that topcodes vary by year. check the [asec codebooks](https://cps.ipums.org/cps/codebooks.shtml) for topcodes applied to each source variable each year. for example, based on the [2019 asec codebook](https://cps.ipums.org/cps/resources/codebooks/cpsmar19.pdf), stampval was crea…",
"name": "",
"is_accepted": true
}
] |
e79a616ee6be945e99cfed8d16682900 | the [proptx99 availability tab](https://usa.ipums.org/usa-action/variables/proptx99#availability_section) shows this variable is not available after 2017.
what’s the reason for this? (is it no longer collected by the census bureau? or does it have a different name?)
i am asking because related variables such as propinsr, taxincl and insincl are available after 2017. | proptx99 after 2017 | ipums.org | 2021.25 | [
{
"text": "the harmonized version of this variable has an error which is currently being fixed by the ipums usa team for 2018 and 2019. it should be released in the next few months. in the meantime you can merge the source variable onto your extract using the files available at [data.census.gov/mdat](http://data.census.gov/mdat). the variab…",
"name": "",
"is_accepted": true
}
] |
5488cdda8d66c5e8b78b7d704778744b | apologies if i am overlooking something here!
i am conducting a detailed analysis of retirement income using the unharmonized asec 20 dataset. part of this requires identifying certain types of survivor’s income by source.
the unharmonized ipums dataset has survivor’s income value by primary/secon… | asec 20 unharmonized survivor's income soruces | ipums.org | 2021.10 | [
{
"text": "these variables will be added to ipums in the next data release, scheduled for this friday. they will be named uh\\_incss1\\_a3 and uh\\_incss2\\_a3, respectively in ipums cps once released.",
"name": "",
"is_accepted": true
}
] |
d5b074df8112997447492d72f174325b | i am looking to distinguish between topcoded and n.i.u. values for the variable incunemp. looking over the documentation, i am confused as to which value is the top code and which is the n.i.u. value. i asked a somewhat similar question [here](https://forum.ipums.org/t/income-top-code-discrepancies-oincbus/3263/2) in regard to the oincbus variable.
when looking at the [do…](https://cps.ipums.org/cps-action/variables/incunemp#codes_section) | n.i.u. value confusion incunemp | ipums.org | 2020.05 | [
{
"text": "the answer is the same as your [previous question](https://forum.ipums.org/t/income-top-code-discrepancies-oincbus/3263/2). the top-code depends on the specific sample year for these variables. for both [incunemp](https://cps.ipums.org/cps-action/variables/incunemp#codes_section), 99997 is the topcode for years 1988-1995 and for other years the topcode is specified in the [topcodes tables](https://cps.ipums.org/cps/topcodes_tables.shtml). also note the following from the topcode page under…",
"name": "",
"is_accepted": true
}
] |
7b1200daf32499775c90bcc12b9ae600 | i have found that 24 of the 2010 acs occupation codes that change between the 2010 and 2018 classification systems are not sampled in the cps after april 2012 (e.g. 2960, 6430, 7930, 8840, 9500). does anyone know why this might be? clearly these occupations are redundant by 2018 (as these codes are … | 2010 occupation codes missing since april 2012 | ipums.org | 2021.39 | [
{
"text": "the cps altered their occupation coding scheme in 2012, which simply grouped a few small occupations together. it’s not very well documented, but there’s evidence of the change by looking at the [cps codebooks](https://cps.ipums.org/cps/codebooks.shtml) starting in may 2012—the appendix listing occupation codes notes “beginning may 2012”. the …",
"name": "",
"is_accepted": true
}
] |
124ee47d4be01eccc172f53b281a42c0 | i am working on a project where i need to derive numbers of children aged three and four by state in different years. for this project, i am using cps asec data. i first calculated the sums of children aged three and four in years 1990, 1991, and 1992 by first identifying the children using the age … | calculating numbers of young children by state using asec weights | ipums.org | 2020.05 | [
{
"text": "the short answer is that the weights in the cps are not constructed to ensure that every age-state combination sums to a specified population control. the cps technical documentation gives much more detail on this. see page 10-3 “ratio estimation” in this technical paper: [https://cps.ipums.org/cps/…](https://cps.ipums.org/cps/resources/cpr/tp63.pdf)",
"name": "",
"is_accepted": true
}
] |
5cff9a0533994b88dac1514fa8f23f88 | hi, i suspect the stata extraction code for earnings is incorrect. i get the following means for weekly and hourly earnings:
. su earnweek*
variable | obs mean std. dev. min max
-------------+---------------------------------------------------------
earnweek … | error in stata code for atus and cps earnings vars? | ipums.org | 2020.16 | [
{
"text": "yes, the ipums online data extract system includes any and all documentation available. the codes tab for each variable (e.g., [earnweek](https://www.atusdata.org/atus-action/variables/earnweek#codes_section) and [hourwage](https://www.atusdata.org/atus-action/variables/hourwage#codes_section)) identifies special codes, such as niu. in earnweek niu==99999.99 and in hourwage niu==99.99.",
"name": "",
"is_accepted": true
}
] |
228c859215e11dd00c2dfbd2b6fdd777 | disclaimer: i am a stats newbie, so apologies for asking perhaps an obvious question…
i am running a regression to see how household water and sanitation facilities affect years of schooling for children. i am uncertain as to whether i have weighted the data correctly using svyset. i have had a loo… | svyset for regression using colombia 2005 census | ipums.org | 2020.29 | [
{
"text": "yes, it seems that you have specified this correctly for individual-level regression analysis. note that there are numerous ways to specify sampling weights in stata and so this is only one approach that you can use. [here is a nice resource for using svy in stata](https://www.stata.com/meeting/canada09/ca09_pitblado_handout.pdf).",
"name": "",
"is_accepted": true
}
] |
d3d8ffaf58b785f8d2421fa2c69ac927 | hi! i’m working with a merged sample of asec and basic monthly samples which includes data from march 1996 - 2006.
i want to create two variables that need to be constructed by using family interrelationships:
a dummy variable which indicates whether anyone in the family of the individual of int… | family members in bad health | ipums.org | 2020.05 | [
{
"text": "okay, yes, i think you are on the right track regarding merging the march basic monthly sample with the asec sample. i just wanted to flag this detail in case there was any confusion.\n\nregarding the difficulties you’ve noted here. on the detail about not wanting to count the “individual of interest,…",
"name": "",
"is_accepted": true
}
] |
b4d84e875750e9a01ed6f9701c26c330 | i am working with the cps asec data and i noticed something unexpected about the ownershp variable:
the case-count view shows a large number of observations (households) with niu status.
[[screen shot 2021-06-25 at 1.18.33 pm]](https://aws1.discourse-cdn.com/business5/uploads/ipums/original/1x/492b04fd6f922641222edd3760eaaec0cfcc986d.png "screen shot 2021-06-25 at 1.18.33 pm")
however, when i inspect the data file included in my extract, i do not… | asec: ownershp - niu | ipums.org | 2021.31 | [
{
"text": "the case counts for household level variables on the ipums website include all households sampled for the cps, whether they were interviewed or not. the non-interviewed households are included in hierarchical data extracts. however they are not included in rectangular extracts, which is the default …",
"name": "",
"is_accepted": true
}
] |
34340edd8f0aadcb8fcb0a1acaebe102 | hi,
i am looking for information on/lists of unincorporated places in historic census data (c. 1920). ideally, geographic coordinates of all unincorporated places. i notice [here](https://nhgis.org/documentation/gis-data/place-points#1900-1920) a reference to lists of unincorporated places that have been integrated into nhgis data. is this available for downloa… | unincorporated places 1920 | ipums.org | 2021.17 | [
{
"text": "we have create a gis file that shows the location of essentially all incorporated and unincorporated places listed in the published census volumes for 1920. unfortunately, the gis file does not differentiate between incorporated and unincorporated places. we simply provide a unique identifier, a sta…",
"name": "",
"is_accepted": true
}
] |
bc74d84090b9bb5963d24636c1f3b020 | i can attach spouse characteristics for my selected sample, but do those partners also show up in the individual data? that is, suppose i have the following situation: individual 1 is married to individual 10 and individual 2 is married to individual 20. is it the case that individuals 10 and 20 are… | married couples in the cps | ipums.org | 2020.45 | [
{
"text": "your first example is correct. if you have not restricted your sample using select cases, no records will be excluded from the data. attach characteristics simply saves you the effort of writing code to attach a spouse’s characteristic to a person record, but it does not remove the spouse’s record f…",
"name": "",
"is_accepted": true
}
] |
834353d034193b6d5c97ace4349a4f83 | hi all,
i’m curious why there are so many missing values for the census tract-level poverty table? i’m using the time series data named [persons* below poverty level in previous year]. in addition, there are in total more than 111,000 listed census tracts in that table, isn’t the number of census t… | census tract poverty missing values | ipums.org | 2021.10 | [
{
"text": "you’ll need data for block group parts, not block groups. block group parts are smaller partitions of block groups. see the documentation i linked to for more info. but then yes, you can use that data to generate 1990 and 2000 poverty data for 2010 tracts!",
"name": "",
"is_accepted": true
}
] |
23258e52ffc3aeb632a648a0d0b2876f | is there a document (or better yet a script) that contains a mapping between all the occupation classification across all changes that the census bureau implemented? that is, something like a table that maps all occupation codes to the classification for all years? as far as i know, the different ma… | converting occupation codes accross years | ipums.org | 2020.34 | [
{
"text": "ipums provides harmonized versions of occupation (and industry) that do just this–coding all the different iterations of the census occupation coding scheme to a single version of it; we generate these by compiling the crosswalks you describe that track changes across two adjacent schemes. see the v…",
"name": "",
"is_accepted": true
}
] |
0cc0dbbe5f5dd142ebcc3132a548987a | i’m using asec data to extract some descriptive analysis of labor market characteristics of native and foreign workers. but stata does not allow asecwt as fweight in the ‘contract’ command since numbers are not integers.
my question is: can i generate a new weight variable by rounding asecwt and th… | asec national level frequencies with stata - fweight or pweight | ipums.org | 2019.47 | [
{
"text": "yes, you can round the weights and the difference will be negligible. if you’re skeptical try a command (like -mean-) that accepts non-integer weights and compare it with the unrounded and rounded weights. here’s one i just did:\n\noriginal weights:\n\n---------------------------------------------------…",
"name": "",
"is_accepted": true
}
] |
95b8b198be5895f31db20cf0e1ebd419 | dear ipums,
thanks for supplying the world with useful data. i frequently work with the asec, but recently downloaded a large chunk of the monthly cps data. i’m running into an issue where the data do not appear to be identified by year, serial, and pernum, or at the very least seem to be scrambled… | data appears to be scrambled | ipums.org | 2020.40 | [
{
"text": "check cpsid and cpsidp to track households and persons across different cps monthly samples.\n\ntheir values from your image indicate that, for example, lines 2 to 4 is the same person which is consistent with the age and gender variable.",
"name": "",
"is_accepted": true
}
] |
c3193abb631b4ee6a0ab8a6c01d9cf72 | i don’t see a time series involving median income and race.
how would you advise to create my own time series based on:
• 1970, 1980, 1990, 2000, 2010
• county level (or metro level)
• median household (or family) income
• black/white preferably non-hispanic black, non-hispanic white
• i had a… | advice on median income for black/white | ipums.org | 2020.50 | [
{
"text": "in nhgis you can use the filters from the [main data finder](https://data2.nhgis.org/main) to enter these criteria and search through available tables.\n\nfor example, i entered the following:\n\ngeographic levels : county\n\nyears : 2010\n\ntopics : race (topic filter), household & family income (topic filter)\n\nthis will create a list …",
"name": "",
"is_accepted": true
}
] |
bc6450a237c41d6b414a47d1cf1ace06 | as i’m working on a a household level analysis that requires ownership, household income, and demographic variables, i’m wondering which variables correspond to the survey year and which to the previous calendar year. hhincome explicitly mentions that it refers to the previous calendar year, but own… | ownershp data refers to current or previous calendar year | ipums.org | 2021.17 | [
{
"text": "unless otherwise specified, most variables in the cps (including ownershp) refer to the week preceding the survey (more detail [here](https://www.bls.gov/cps/definitions.htm#refweek)). if there is a specific reference period that’s different from this, this will be included in the variable description, as you saw for hhincome. as a rule of thumb, va…",
"name": "",
"is_accepted": true
}
] |
0116eb5e26e23a6a42d07b030c6283b4 | is there any ready-made table containing land area for u.s. counties? nhgis provides shapefiles but i cannot find any table. | county land area | ipums.org | 2020.29 | [
{
"text": "nhgis doesn’t have a data table with land area for us counties, but the county shapefiles from 2010 to the present do provide that information. there is a field called “arealand” in the shapefile, and it contains the land area for each county (in square meters). there is also a field called “areawat…",
"name": "",
"is_accepted": true
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.