Search Results

Search found 70288 results on 2812 pages for 'custom data generator'.

Page 1605/2812 | < Previous Page | 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612  | Next Page >

  • How can I Fix the CD/DVD refresh for the media drive so that Ubuntu 10.10 recognize it automatically?

    - by Denja
    Hello Community, I have recently installed Ubuntu 10.10 and it seems to work flawlessly. But today I tried to check some CDs and I was surprised that Ubuntu doesn't automatically refresh the CD/DVDs inserted in my media drive. I had to eject manually the CD/DVD from Ubuntu and then Insert the new CD and only then Ubuntu reads the data in the newly inserted CD/DVD. How can I fix the CD/DVD refresh for the media drive so that Ubuntu 10.10 recognize it automatically?

    Read the article

  • The IT Security Bubble Has Popped

    The blank check IT security enjoyed has been pulled off the table as businesses reassess how much they spend to stay safe. Need some data on the mood? McAfee reports nearly three-quarters of SMBs have cut security spending, even though 71 percent believe a serious attack would put them out of business.

    Read the article

  • How come a keyword with 46 local monthly searches get 150 local monthly impressions?

    - by Geno Thampi
    I am doing a keyword analysis by correlating data between Google ANalytics and Google Adwords Keyword tool. So here is the confusion I check the local monthly searches (Sweden) for "SEO Packages" using the adwords keyword tool and it shows: 46 monthly searches Now I come back to GA and check the impressions for "SEO Packages" that we got from Sweden and the value is: 150 impressions. So basically: Local monthly searches in Sweden: 46 Monthly impression we got from Sweden: 150 How come we get 150 impressions out of only 150 searches?

    Read the article

  • TechEd 2014 Day 3

    - by John Paul Cook
    There is some confusion about durability of data stored in SQL Server in-memory tables, so some review of the concepts is appropriate. The in-memory option is enabled at the database level. Enabling it at the database level only gives you the option to specify the in-memory feature on a table by table basis. No existing tables or new tables will by default become in-memory tables when you enable the feature at the database level. If you choose to make a table an in-memory table, by default it is...(read more)

    Read the article

  • What is best practice for search engines when a website is under maintenance?

    - by jamescridland
    I need around a week to transition a heavily data-driven website from one back end to another. During that time I do plan to attempt to keep some pages live, but they won't all work well or look brilliant. Some pages won't work at all. What is the best way to ensure I don't scare Google? Should I hide everything from robots.txt, or mark everything that doesn't work as "503", or are there other things that I should be considering?

    Read the article

  • How To Be Your Own Personal Clone Army (With a Little Photoshop)

    - by Eric Z Goodnight
    Maybe you’ve always wanted more of yourself. Or maybe you’ve always thought you could be your own best friend! Regardless of your reasons, here’s how to duplicate yourself with some clever photograph tricks and either Photoshop or GIMP. How To Be Your Own Personal Clone Army (With a Little Photoshop) How To Properly Scan a Photograph (And Get An Even Better Image) The HTG Guide to Hiding Your Data in a TrueCrypt Hidden Volume

    Read the article

  • Looking for a web service for students tracking

    - by shannoga
    I am working Voluntary association with a low budget. They asked me to build a tracking system for the students they work with. It is fairly simple, it needs to let them store data on the student personal details and grades and have the ability to get reports and charts on the students achievements. Since their budget is low I thought looking fro a web service that can feet their needs. Any Idea's?

    Read the article

  • core.* files eating up server space (~50MB)

    - by skytreader
    I'm renting server space from someone and, upon logging in my control panel after quite sometime, noticed an abnormal spike (~50MB) in the disk usage. Upon investigating, I found a lot of core.* files scattered around my public_html directory. Each one is more than 5MB in size but no more than 6MB. The * part is all numbers (in programming regex, that should be core\.\d+). I downloaded one and checked the contents. There was a lot of balderdash characters (NUL mostly, but also a scattering of ETB, ETX, STX) but there's this block of readable text which says: This text is part of the internal format of your mail folder, and is not a real message. It is created automatically by the mail system software. If deleted, important folder data will be lost, and it will be re-created with the data reset to initial values. Pretty self-explanatory. A few blocks above the text are some more readable messages that look like logs but is sandwiched in between non printable characters. I've extracted some below. Scan not valid for mh mailboxes Bogus character 0x%x in news state Can't rewrite news state %.80s Error closing backup news state %.80s No state for newsgroup %.80s found Now, a few concerns: Am I under attack? The messages seem to be about my webmail but I don't use my personal webmail that much---only for a vanity email address and an inbox for an outdated comments system. However, lately, I seem to notice a spike in the spam for my vanity mail. (Note: the comments system is covered by a captcha but every now and then some get through. My vanity email has a spam filter but it isn't as good as I'd like). Next, if this is a feature, can I turn it off? Is it advisable to? I've only 150MB so you see why I'm fretting over a 50MB spike. Some final details: my only server-side scripts are in PHP. The directory which accumulated the most number of these core files is the one containing the Wordpress-managed subdomain of my site. I manage my server through CPanel. Lastly, I decided to delete this files and after some checking nothing seems amiss in my websites nor in my mail. They are indeed the ones responsible for the ~50MB spike as my disk space usage is back to expected.

    Read the article

  • Capture a Query Executed By An Application Or User Against a SQL Server Database in Less Than a Minute

    - by Compudicted
    At times a Database Administrator, or even a developer is required to wear a spy’s hat. This necessity oftentimes is dictated by a need to take a glimpse into a black-box application for reasons varying from a performance issue to an unauthorized access to data or resources, or as in my most recent case, a closed source custom application that was abandoned by a deserted contractor without source code. It may not be news or unknown to most IT people that SQL Server has always provided means of back-door access to everything connecting to its database. This indispensible tool is SQL Server Profiler. This “gem” is always quietly sitting in the Start – Programs – SQL Server <product version> – Performance Tools folder (yes, it is for performance analysis mostly, but not limited to) ready to help you! So, to the action, let’s start it up. Once ready click on the File – New Trace button, or using Ctrl-N with your keyboard. The standard connection dialog you have seen in SSMS comes up where you connect the standard way: One side note here, you will be able to connect only if your account belongs to the sysadmin or alter trace fixed server role. Upon a successful connection you must be able to see this initial dialog: At this stage I will give a hint: you will have a wide variety of predefined templates: But to shorten your time to results you would need to opt for using the TSQL_Grouped template. Now you need to set it up. In some cases, you will know the principal’s login name (account) that needs to be monitored in advance, and in some (like in mine), you will not. But it is VERY helpful to monitor just a particular account to minimize the amount of results returned. So if you know it you can already go to the Event Section tab, then click the Column Filters button which would bring a dialog below where you key in the account being monitored without any mask (or whildcard):  If you do not know the principal name then you will need to poke around and look around for things like a config file where (typically!) the connection string is fully exposed. That was the case in my situation, an application had an app.config (XML) file with the connection string in it not encrypted: This made my endeavor very easy. So after I entered the account to monitor I clicked on Run button and also started my black-box application. Voilà, in a under a minute of time I had the SQL statement captured:

    Read the article

  • Export to XML Using SSIS

    Exporting data to XML format using SSIS initially seems like it should be straight forward – just dump it in a flat file and give it a name xml type, however SSIS has no XML destination just an XML source. I had no luck dumping the XML in flat file. My solution was to use a script task which worked well. Check SQL Server performance at a glanceWe consulted 1000 SQL Server professionals to make SQL Monitor’s UI as clear as possible. Start monitoring with a free trial.

    Read the article

  • How would you price a dynamic real estate property management website? [closed]

    - by user1217550
    Imagine, hypothetically of course, that you are being commissioned to develop a full-fledged real estate website that includes: 1) a search engine with ajax/json autofill, 2) google maps and geolocation integration, google streetview, 3) user registration, login and account management 4) administrative panels to control data input 5) search results page 6) user statistics 7) property inquiry to allow internal messaging between users How much would you charge? Suppose you are developing the most advanced and specific system in PhP/MySQL, and your total development time is roughly 1500 hours? Any suggestions?

    Read the article

  • Practical PowerShell for SQL Server Developers and DBAs – Part 1

    There is a lot of confusion amongst DBAs about using PowerShell due to existence the deprecated SQLPS mini-shell of SSMS and the newer SQLPS module. In a two-part article and wallchart, Michael Sorens explains how to install it, what it is, and some of the excellent things it has to offer. Compress live data by 73% Red Gate's SQL Storage Compress reduces the size of live SQL Server databases, saving you disk space and storage costs. Learn more.

    Read the article

  • Defining and selecting an area from an image

    - by meth0d_
    I want to create a game, where the world is loaded from an image file, much like Paradox Interactive does it for their games. If I have this image: Then the red, green, blue, cyan, magenta, white, black and grey should be different provinces. I know how to loop through them, and check if it's a new province, the problem is that I don't know how to define the region of the province for selection: I don't know how I can load in data, to make sure you can click anywhere on that province, and make sure it gets selected.

    Read the article

  • linode.com/slicehost.com/vps.net what to chose? [closed]

    - by Guy
    Possible Duplicate: How to find web hosting that meets my requirements? I am looking for a new VPS for http://hotelpublisher.com. At the moment it is either linode.com, slicehost.com or vps.net (alternatives are welcome). Since I already use Google cloud to deliver data, my priority is ram/cpu/reliability/price. Can anyone advice which of the VPS providers is the best in their opinion and why?

    Read the article

  • SQL Server Unit Testing with tSQLt

    When one considers the amount of time and effort that Unit Testing consumes for the Database Developer, is surprising how few good SQL Server Test frameworks are around. tSQLt , which is open source and free to use, is one of the frameworks that provide a simple way to populate a table with test data as part of the unit test, and check the results with what should be expected. Sebastian and Dennis, who created tSQLt, explain.

    Read the article

  • SQL Server 2008 SP3 and HP Database Enterprise Appliance

    The list of updated features begins with an overall improved upgrading experience from SQL Server's older versions to the latest SQL Server 2008 SP3 edition. Microsoft has implemented measures that provide enhanced performance and reliability while making the upgrade as well. Logs in SQL Server Integration Services will display the aggregate number of rows transmitted in Data Flows, and enabling the Shrink Database option will produce enhanced warning messages when maintenance plan development takes place. Microsoft's blog post continues with the list of SP3 upgrades, noting that users sho...

    Read the article

  • How to: using ASP.NET 4 QueryExtender control

    This article demonstrates how to use newly added QueryExtender control in ASP.NET 4 to filter data dynamically on a web page using declarative syntax...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Google Chrome Loses ASP.NET Sessions - Need FavIcon

    - by nannette
    I had programmed a brilliant web page in ASP.NET 4.0 and lo and behold, this one page lost its sessions in Google Chrome. I could run it locally in debug and could not reproduce the issues in Chrome. Didn't happen in IE or Firefox, only Chrome on the published server. I finally found in a forum where someone mentioned that Google Chrome looks for favicons and if it doesn't find one it will throw a 302 redirect and kill the session. http://stackoverflow.com/questions/8247842/session-data-lost-in-chrome...(read more)

    Read the article

  • SQL Server 2012 Integration Services - Package and Project Parameters

    In SQL Server 2012, Microsoft introduced SQL Server Data Tools to accommodate the dynamic nature of SSIS constructs in the form of package and project parameters. This approach lets you combine multi-package projects into a single unit, eliminating the possibility of breaking dependencies between parent and child packages during subsequent deployments. Get smart with SQL Backup ProGet faster, smaller backups with integrated verification.Quickly and easily DBCC CHECKDB your backups. Learn more.

    Read the article

  • Common database deployment blockers and Continuous Delivery headaches

    Deployability is now a first class concern for databases, so why isn’t it as easy as it should be? Matthew Skelton explores seven of the most common challenges which will bring your database deployments to their knees. Get alerts within 15 seconds of SQL Server issuesSQL Monitor checks performance data every 15 seconds, so you can fix issues before your users even notice them. Start monitoring with a free trial.

    Read the article

  • Oracle Days 2013 in EMEA Are Coming Soon: Invite Your Customers!

    - by Javier Puerta
    Oracle Days will again be hosted across EMEA this October and November (schedule here). By attending an Oracle Day, your customers can: Hear the new announcements from Oracle OpenWorld See customer case studies showing innovation in practice. Discuss key issues for business and IT executives in cloud, mobile, social, big data, The Internet of Things Network with peers who are facing the same challenges Meet Oracle experts and watch live demos of new products  Promoting Oracle Day to Your Customers Follow the Oracle Day schedule in EMEA Direct your customers to the Oracle Day 2013 video on Oracle.com

    Read the article

  • Sampling SQL server batch activity

    - by extended_events
    Recently I was troubleshooting a performance issue on an internal tracking workload and needed to collect some very low level events over a period of 3-4 hours.  During analysis of the data I found that a common pattern I was using was to find a batch with a duration that was longer than average and follow all the events it produced.  This pattern got me thinking that I was discarding a substantial amount of event data that had been collected, and that it would be great to be able to reduce the collection overhead on the server if I could still get all activity from some batches. In the past I’ve used a sampling technique based on the counter predicate to build a baseline of overall activity (see Mikes post here).  This isn’t exactly what I want though as there would certainly be events from a particular batch that wouldn’t pass the predicate.  What I need is a way to identify streams of work and select say one in ten of them to watch, and sql server provides just such a mechanism: session_id.  Session_id is a server assigned integer that is bound to a connection at login and lasts until logout.  So by combining the session_id predicate source and the divides_by_uint64 predicate comparator we can limit collection, and still get all the events in batches for investigation. CREATE EVENT SESSION session_10_percent ON SERVER ADD EVENT sqlserver.sql_statement_starting(     WHERE (package0.divides_by_uint64(sqlserver.session_id,10))), ADD EVENT sqlos.wait_info (        WHERE (package0.divides_by_uint64(sqlserver.session_id,10))), ADD EVENT sqlos.wait_info_external (        WHERE (package0.divides_by_uint64(sqlserver.session_id,10))), ADD EVENT sqlserver.sql_statement_completed(     WHERE (package0.divides_by_uint64(sqlserver.session_id,10))) ADD TARGET ring_buffer WITH (MAX_DISPATCH_LATENCY=30 SECONDS,TRACK_CAUSALITY=ON) GO   There we go; event collection is reduced while still providing enough information to find the root of the problem.  By the way the performance issue turned out to be an IO issue, and the session definition above was more than enough to show long waits on PAGEIOLATCH*.        

    Read the article

< Previous Page | 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612  | Next Page >