Search Results

Search found 16838 results on 674 pages for 'writing patterns dita cms'.

Page 374/674 | < Previous Page | 370 371 372 373 374 375 376 377 378 379 380 381  | Next Page >

  • Text after Control Sequence

    - by SPAM SPAM SPAM SPAM
    I am trying to parse the output of a command that expects to be writing to the screen. It has data separated by move-to-origin control sequences (for the VT220, ESC[1;1H). I only need the last part (i.e. after the last move-to-origin). I have tried doing this multiple ways (primarily awk and sed), but the problem is always that parts of the control sequence have special meaning (to the program, not just to the shell), and I cannot quote them when I substitute tput's output. Any suggestions?

    Read the article

  • compromised site

    - by pinniger
    So, I have a web site that has been compromised twice in two weeks. every index.php and .js file gets a script injecting into the source code of the file. The problem is that I have no idea how they're doing it. I've seen this done via sql injection before, but I don't know how they are actually writing to the file. I've dug through the Apache logs but didn't find anything interesting. The site is built using the cakephp framework on a godaddy shared server. Anybody know what secturity settings or log files to check to see how they are doing this?

    Read the article

  • What are the pros & cons of these MySQL engines for OLTP -- XtraDB, PBXT, or TokuDB?

    - by Continuation
    I'm working on a social website with an approximate read/write split of 90/10. Trying to decide on a MySQL engine. The ones I'm interested in are: XtraDB PBXT TokuDB What are the pros and cons of them for my use case? A few specific questions: PBXT uses log-based structure that avoids double-writes. It sounds very elegant, but the benchmark I've seen doesn't show any/much advantages over XtraDB. Do you have any experience with PBXT/XtraDB you can share? TokuDB sounds VERY interesting. But all the benchmarks I've seen are about single-threaded bulk inserts - inserting 100M rows for example. that's not very relevant for OLTP. What about its performance with large number of concurrent threads writing and reading at the same time? Anyone has tried that?

    Read the article

  • Looking for "pad" for my keyboard + mouse in a multi-monitor setup

    - by Borek
    I've got 2 large monitors. My keyboard and mouse are in the middle of my desk which is appropriate for work on both monitors but when I want to focus on one of them only (say that I'm writing in MS Word on the right monitor), I'd like to move my keyboard and mouse to the right. I can do that manually but because of all the wires and other things, it is quite inconvenient when done frequently. I'm looking for something that I would put under both my keyboard and my mouse and that would slide to one side or the other easily. Is there anything like that available on the market?

    Read the article

  • Encrypted directory makes file operations for whole disk very slow

    - by user1566277
    I am running an arm GNU/Linux and I have a SD-Card with three partitions on it. On one of the Partition I create an encfs file and then mount it on a directory which is in another partition to make that directory encrypted. Works fine. But now the writing speed on all the partitions are reduced drastically. I can understand that it should be slow for encrypted directory but why the its reducing write speed for all the partitions. E.g., if do not mount the encrypted directory 20MB is transferred in 2 Sec. roughly but with the encrypted directory mounted its like 20 Seconds for same file. I am using LUKS and all the partitons are ext3 except for the directory where /dev/mapper/encfs is mounted as type ext2. Any hints?

    Read the article

  • File report ignoring NTFS permissions

    - by Edulin
    I need to inventory based on file age a huge NAS. Some folders are restricted, and only users or groups have rights. As you probably know Treesize is the best solution, the only problem I found is that if I want to script it it only takes a screenshot instead of writing the information to a text file. (If done through the GUI I'm allowed to copy/paste into text, but I need to script it, several shares). Robocopy /l lists the files but I also need to know the dates (last access / last modify) I saw a kernel driver by hobeanu to bypass ntfs permissions. But do you know any easier way to bypass ntfs permissions? or any other application that allows this type of request and save the information to a text file. (scriptable) Thanks

    Read the article

  • PING through batch file (looping problem)

    - by pradeetp
    I created a .BAT file in Windows 7 that has the following lines: PING XXX.XXX.XXX.XXX -t (XXX replaces the actual IP number). However, when I double click on this batch file, I can see the PING command repeatedly being executed in a loop. I even tried to rename the .BAT to .CMD but the result is the same. I want to avoid writing PING command through the Command Prompt, which is why I created the batch file. I don't know why the PING command is being continuously called when the same statement is put in a batch file.

    Read the article

  • How Do I Migrate 100 DBs From One MS-SQL 2008 Server To Another? (looking for automation)

    - by jc4rp3nt3r
    Let me start by saying that I am not a DBA, but I am in a position where I am responsible for moving just under 100 MS-SQL 2008 DBs from our current development server, to a new/better/faster development server. As this is just a local dev server, temporary downtime is acceptable, but I am looking for a way to move all of the databases (preferably in bulk). I know that I could take a bak of each, and restore it on the new server, but given the volume of DBs, I am looking for a more efficient way. I am not opposed to learning a new piece of software, writing code or any other requirement, so long as it speeds up the process.

    Read the article

  • Computer restarts infinitely after Bios/Motherboard screen

    - by user13176
    Like every morning I turn on my computer (which is kinda new and only couple months old) and I never had any issue with it until now. It starts booting, shows that very first screen which has some motherboard picture and stuff and then it show which buttons to press to enter BIOS and etc. Then it suddenly restarted and after that just kept restarting every 3-5 seconds forever. What could that be? I unplugged power cable, checked all the cables, moved a case, it had some wires stuck under it, blew some dust (but didn't open the case yet) and I tried to boot again and it booted successfully and I am writing this issue now. My specs: i3-3220 3.30Ghz 8GB DDR3 RAM 1TB HDD GeForce GTX 650 Windows 8 x64 I also recently upgraded my security software (that's the most recent thing I did) and it includes ZoneAlarm firewall, Comodo Antivirus and Spybot Search & Destroy. But maybe it doesn't have anything to do with this. Cheers

    Read the article

  • Debian SMB share having permission issues in windows 7

    - by xxpor
    Hi Everyone, I setup a Debian squeeze server with samba. I then shared my /media directory with the following configuration: [media] comment=Hard Drives read only = no locking = no path = /media guest ok = yes browsable = yes directory mask = 0777 When the drives are mapped in Windows 7, the user can write to all of the subdirectories of media (sdb1, sdc1, etc), but cannot write to any folders that they create themselves in the subdirectories of media. For example, if the user mapped /media/sdb1 to Z:, and then creates a folder Z:\test, the folder is created successfully, but no files can be written to Z:\test. If the user ssh's into the server, they have no problems writing to these directories. I have included the screenshots, in order, of what happens on Windows. This samba share is mounted with ntfs-3g, if it makes a difference. Screenshots are here

    Read the article

  • Using (embedding?) wireshark in a C application for sniffing

    - by happy_emi
    I'm writing a C/C++ application which needs (among other things) to sniff packets and save the output in a file. This file will be read and processed by wireshark after a few days, using a LUA script to do some other stuff. My question is for the sniffing part which must be provided by my application. I can see two ways to do this: 1) Fork the wireshark process in background (of course using the command line version) 2) Using wireshark as library: no forking but include stuff like "wireshark.h" and link against libwireshark.so, thus using function calls to do the sniffing. So far I haven't found any documentation about #2 and it seems that #1 is the "right way" to embed sniffing capabilities in my code. Do you think I'm doing he right thing? Does wireshark allow embedding as a library?

    Read the article

  • Running a cronjob

    - by Ed01
    've been puzzling over cronjobs for the last few hours. I've read documentation and examples. I understand the basics and concepts, but haven't gotten anything to work. So I would appreciate some help with this total noob dilemma. The ultimate goal is to schedule the execution of a django function every day. Before I get that far, I want to know that I can schedule any old script to run, first once, then on a regular basis. So I want to: 1) Write a simple script (perhaps a bash script) that will allow me to determine that yes, it did indeed run successfully, or that it failed. 2) schedule this script to run at the top of the hour I tried writing a bash script that simple output some text to the terminal: #!/bin/bash echo "The script ran" Then I dropped this into a .txt file MAILTO = *****.******@gmail.com 05 * * * * /home/vadmin/development/test.sh But nothing happened. I'm sure I did many things wrong. Where do I start to fix all of this?

    Read the article

  • Does programmable hardware exist to allow hardware to be programmed by computers?

    - by agentbanks217
    I am a programmer and I have never really dealt with the hardware of anything, only software. I want to start building things that I can control from my computer using programming. My question is are there such devices on the market that have a programmable interface or API? For example, I want to build an automated window blinds opening/closing device, and I would like to be able to control it from my computer e.g. writing an app or some code to schedule them when to open and close. I would like to know if there are any devices that can be programmed to do that (the computer part)?

    Read the article

  • Samba Share writable

    - by Chris
    I have had a problem writing to a Samba share. I believe this person has the answer, but I do not know how to do this, does someone know how to do this? Thank you very much, On the Samba server, you need to ensure that the nobody user has write permissions to /Windows_Backups/DC. You're forcing everyone to be impersonated by the nobody account, so that account will need file-level permissions on that share directory. Samba will respect local permissions when figuring out who can write where, in this case it is somewhat like Windows.

    Read the article

  • Scaling web application with SQL Server 2008 database

    - by John
    I have a database which has 90% of read only tables. 10% of the tables has writable data. We need to scale the ASP.NET application.We need to add more users who will not be writing to the database. We are thinking of adding another server and routing the users who need read only access to that server. Is there a way to replicate just some tables to another database server. Since the 90% of data doesnt change, we don't want to setup any full database replication. Please advise.

    Read the article

  • Does a hard drive degrade if we always write to the same directory?

    - by code-gijoe
    I have a very heavy I/O application that is constantly receiving data through the network and writing to a specific directory on the HDD. Then, the application need to load the files from that specific place. One of my clients has been experiencing slowness and when I try to access the directory it takes quite long before I can see the content. My gut feeling is the HDD is degrading due to high I/O for a couple of years and I'm thinking of changing the HDD. Is there a benefit to write to multiple directories instead of using always the same? BTW he is using Windows XP.

    Read the article

  • Software update shows "P??roblem with the system installer tool" error in OS X Lion

    - by Elnaz Shahmehr
    I can't update my system. Does anyone know what the problem is? sudo softwareupdate -i -a Password: Software Update Tool Copyright 2002-2010 Apple Downloading Digital Camera Raw Compatibility Update Waiting to install Digital Camera Raw Compatibility Update Downloading Java for OS X 2012-002 Verifying Java for OS X 2012-002 Waiting to install Java for OS X 2012-002 Downloading iTunes Waiting to install iTunes Downloading Safari Verifying Safari Waiting to install Safari Checking packages… Installing Waiting for other installations to complete… Validating packages… Writing files… Package failed: There was a problem with the system installer tool. Package failed: There was a problem with the system installer tool. Package failed: There was a problem with the system installer tool. Package failed: There was a problem with the system installer tool. Done.

    Read the article

  • Werid formating in Word 2010

    - by Stat-R
    A few months ago, while writing a paper, I copied some paragraphs created in another computer into a different computer. I guess the formatting was different. Please see the following image: I noticed that a strange formatting has also been imported. I thought it would go away when I select all and choose a format. But the problem did not go away. Now, when I am trying to finish the paper, the weird formatting still remaining. Does anyone have any solution? Also, how to make sure that when we copy something from a file with different Styles, we retain the destination style definitions. EDIT I would prefer a solution where I do not have to re-do the formatting manually.

    Read the article

  • Weird formatting in Word 2010

    - by Stat-R
    A few months ago while writing a paper, I copied some paragraphs created in one computer to a different computer. I guess the formatting was different. Please see the following image: I noticed that a strange formatting has also been imported. I thought it would go away when I select all and choose a format. But the problem did not go away. Now, when I am trying to finish the paper, the weird formatting still remains. Does anyone have any solution? Also, how to make sure that when we copy something from a file with different Styles, we retain the destination style definitions? EDIT I would prefer a solution where I do not have to re-do the formatting manually.

    Read the article

  • SQL SERVER – Generate Report for Index Physical Statistics – SSMS

    - by pinaldave
    Few days ago, I wrote about SQL SERVER – Out of the Box – Activity and Performance Reports from SSSMS (Link). A user asked me a question regarding if we can use similar reports to get the detail about Indexes. Yes, it is possible to do the same. There are similar type of reports are available at Database level, just like those available at the Server Instance level. You can right click on Database name and click Reports. Under Standard Reports, you will find following reports. Disk Usage Disk Usage by Top Tables Disk Usage by Table Disk Usage by Partition Backup and Restore Events All Transactions All Blocking Transactions Top Transactions by Age Top Transactions by Blocked Transactions Count Top Transactions by Locks Count Resource Locking Statistics by Objects Object Execute Statistics Database Consistency history Index Usage Statistics Index Physical Statistics Schema Change history User Statistics Select the Reports with name Index Physical Statistics. Once click, a report containing all the index names along with other information related to index will be visible, e.g. Index Type and number of partitions. One column that caught my interest was Operation Recommended. In some place, it suggested that index needs to be rebuilt. It is also possible to click and expand the column of partitions and see additional details about index as well. DBA and Developers who just want to have idea about how your index is and its physical statistics can use this tool. Click to Enlarge Note: Please note that I will rebuild my indexes just because this report is recommending it. There are many other parameters you need to consider before rebuilding indexes. However, this tool gives you the accurate stats of your index and it can be right away exported to Excel or PDF writing by clicking on the report. Reference : Pinal Dave (http://blog.SQLAuthority.com) Filed under: Pinal Dave, SQL, SQL Authority, SQL Index, SQL Optimization, SQL Performance, SQL Query, SQL Server, SQL Server Management Studio, SQL Tips and Tricks, SQL Utility, T SQL, Technology

    Read the article

  • OpenGL extension vs OpenGL core

    - by user209347
    I was doubting: I'm writing a cross-platform engine OpenGL C++, I figured out windows forces the developers to access OpenGL features above 1.1 through extensions. Now the thing is, on Linux, I know that I can directly access functions if the version supports it through glext.h and opengl version. The problem is that if on Linux, the core doesn't support it, is it possible there is an extensions that supports the same functionality, in my case vertex buffer objects? I'm doing something like this: Windows: (hashdeck) define glFunction functionpointer_to_the_extension (apparently the layout changes font size if I use #) Linux: Since glext already defined glFunction, I can write in client code glFunction, and compile it both on Windows AND Linux without changing a single line in my client code using the engine (my goal). Now the thing is, I saw a tutorial use only the extension on Linux, and not checking for the opengl implementation version. If the functionality is available in the core, is it also available as extension (VBO's e.g.)? Or is an extension something you never know is available? I want to write an engine that gets all the possibilities on hardware, so I need to check (on Linux) for extensions as well as core version for possible functionality implementation.

    Read the article

  • Speaker at developer conferences and user group meetings

    Catching up on a couple of sessions I did in the past. This article gives an overview of some of my activities. Mainly at the annual German Visual FoxPro Developer Conference also known as SQL-Server & ASP.NET Conference in Frankfurt. The below listed entries are excerpts from the original Conference Coverage documents you'll find on UniversalThread. German Visual FoxPro Developer Conference 2002 (1 session - Vendor session about Active FoxPro Pages 3.0) German Visual FoxPro Developer Conference 2003 (2.5 sessions - Visual FoxPro running on Linux) German Visual FoxPro Developer Conference 2004 (4 sessions - 2x Active FoxPro Pages, VFP on Linux, and VFP using additional databases) German Visual FoxPro Developer Conference 2005 (4 sessions - RegEx, XML, XSLT, and using free (as in beer) development tools) German Visual FoxPro Developer Conference 2006 (3 sessions - .NET interop via COM, writing own CLR host in VFP, and Active FoxPro Pages) Furthermore, I did a couple of (hopefully) interesting sessions at various user group meetings in Speyer and Stuttgart. A more comprehensive list is available under Presentations (in German language). And last but not least, back in May 2005 Microsoft Germany invited me to host a WebCast for MSDN on how to use 'Visual FoxPro mit Visual Studio 2005'. Unfortunately, I was too unexperienced and too nervous (first time ever), we experienced technical issues with the microphone, and the obviously low quality of recording demanded to replace it by a whole series on Visual FoxPro 9.0. The webcast covered the same topics I already described in other articles here on my blog. Despite the desaster I'd like to thank Ralf Westphal for his kind words afterwards - I really felt bad. Eventually, you might ask yourself why it stopped by the end of 2006... Well, new chapter in my life: Mauritius!

    Read the article

  • SQLAuthority News – Three Posts on Reporting – T-SQL Tuesday #005

    - by pinaldave
    If you are following my blog, you already know that I am more of “T-SQL and Performance Tuning” type of person. I do have a good understanding of Business Intelligence suit and I also do certain training sessions on the same subject. When I was writing the blog post for T-SQL Tuesday #005 – Reporting, I realized that I have written a post that clearly explains how to generate reports using SQL Server Management Studio. Here is a quick recap on how one can use SSMS and out-of-the-box reports which can help many developers. Please note that they can be resource-intensive as well, so please use SSMS carefully. SQL SERVER – Generate Report for Index Physical Statistics – SSMS SQL SERVER – Out of the Box – Activity and Performance Reports from SSSMS SQL SERVER – Configure Management Data Collection in Quick Steps – T-SQL Tuesday #005 Junior developers and DBA can use these reports right away and can also start learning and exploring most database performance issues with the help of Sr. DBAs. Reference: Pinal Dave (http://blog.SQLAuthority.com) Filed under: Pinal Dave, SQL, SQL Authority, SQL Performance, SQL Query, SQL Server, SQL Server Management Studio, SQL Tips and Tricks, T SQL, Technology Tagged: SQL Reporting, SQL Reports

    Read the article

  • SQLAuthority News – Statistics Used by the Query Optimizer in Microsoft SQL Server 2008 – Microsoft Whitepaper

    - by pinaldave
    I recently presented session on Statistics and Best Practices in Virtual Tech Days on Nov 22, 2010. The sessions was very popular and I got many questions right after the sessions. The number question I had received was where everybody can get the further information. I am very much happy that my sessions created some curiosity for one of the most important feature of the SQL Server. Statistics are the heart of the SQL Server. Microsoft has published a white paper on the subject how statistics are useful to Query Optimizer. Here is the abstract of the same white paper from Microsoft. Statistics Used by the Query Optimizer in Microsoft SQL Server 2008 Writer: Eric N. Hanson and Yavor Angelov Microsoft SQL Server 2008 collects statistical information about indexes and column data stored in the database. These statistics are used by the SQL Server query optimizer to choose the most efficient plan for retrieving or updating data. This paper describes what data is collected, where it is stored, and which commands create, update, and delete statistics. By default, SQL Server 2008 also creates and updates statistics automatically, when such an operation is considered to be useful. This paper also outlines how these defaults can be changed on different levels (column, table, and database). In addition, it presents how certain query language features, such as Transact-SQL variables, interact with use of statistics by the optimizer, and it provides guidance for using these features when writing queries so you can obtain good query performance. Link to white paper Statistics Used by the Query Optimizer in Microsoft SQL Server 2008 ?Reference: Pinal Dave (http://blog.SQLAuthority.com)   Filed under: Pinal Dave, SQL, SQL Authority, SQL Documentation, SQL Download, SQL Query, SQL Scripts, SQL Server, SQL Tips and Tricks, SQL White Papers, SQLAuthority News, T SQL, Technology

    Read the article

  • SQL SERVER – Configure Management Data Collection in Quick Steps – T-SQL Tuesday #005

    - by pinaldave
    This article was written as a response to T-SQL Tuesday #005 – Reporting. The three most important components of any computer and server are the CPU, Memory, and Hard disk specification. This post talks about  how to get more details about these three most important components using the Management Data Collection. Management Data Collection generates the reports for the three said components by default. Configuring Data Collection is a very easy task and can be done very quickly. Please note: There are many different ways to get reports generated for CPU, Memory and IO. You can use DMVs, Extended Events as well Perfmon to trace the data. Keeping the T-SQL Tuesday subject of reporting this post is created to give visual tutorial to quickly configure Data Collection and generate Reports. From Book On-Line: The data collector is a core component of the Data Collection platform for SQL Server 2008 and the tools that are provided by SQL Server. The data collector provides one central point for data collection across your database servers and applications. This collection point can obtain data from a variety of sources and is not limited to performance data, unlike SQL Trace. Let us go over the visual tutorial on how quickly Data Collection can be configured. Expand the management node under the main server node and follow the direction in the pictures. This reports can be exported to PDF as well Excel by writing clicking on reports. Now let us see more additional screenshots of the reports. The reports are very self-explanatory  but can be drilled down to get further details. Click on the image to make it larger. Well, as we can see, it is very easy to configure and utilize this tool. Do you use this tool in your organization? Reference: Pinal Dave (http://blog.SQLAuthority.com) Filed under: Pinal Dave, SQL, SQL Authority, SQL Performance, SQL Query, SQL Server, SQL Tips and Tricks, T SQL, Technology Tagged: SQL Reporting, SQL Reports

    Read the article

< Previous Page | 370 371 372 373 374 375 376 377 378 379 380 381  | Next Page >