Search Results

Search found 40168 results on 1607 pages for 'text processing'.

Page 234/1607 | < Previous Page | 230 231 232 233 234 235 236 237 238 239 240 241  | Next Page >

  • Create comma seperated values in perl

    - by Mike
    Let's say I have a list of elements @list=(1,2,3); #desired output 1,2,3 And I want to print them as comma seperated values. And most importantly, I do not want the last element to have a comma after it. What is the cleanest way to do this in Perl?

    Read the article

  • Perl, search array

    - by Mike
    What is the smartest way of searching through an array of strings for a matching string in Perl? One caveat, I would like the search to be case-insensitive so "aAa" would be in ("aaa","bbb")

    Read the article

  • python extract from switch output

    - by household2
    I have some info back from a LAN switch as below Vlan 1 is administratively down, line protocol is down Vlan 2 is up, line protocol is up Helper address is 192.168.0.2 Vlan 3 is up, line protocol is up Helper address is not set Vlan 4 is up, line protocol is up Helper address is 192.168.0.2 Vlan 5 is down, line protocol is down Helper address is 192.168.0.2 Vlan 6 is down, line protocol is down Helper address is not set Helper address is not set And the output I'm trying for is Vlan 1,admin down,n/a Vlan 2,up/up, 192.168.0.2 Vlan 3, up/up, not set Vlan 4, up/up, 192.168.0.2 Vlan 5, down/down, 192.168.0.2 Vlan 6, down/down, not set So the helper isn't always there (line 1) sometimes it's set sometimes it isn't, sometimes there are two lines (last Vlan - I only need 1) and the Vlan can have states of admin down, up/up, up/down (not here) and down down. So using Python and pexpect I can get the above output, but I'm having difficulty parsing out the consecutive lines. I've tried enumerate and then use key+1 for the next line, but the fact that there can be 0,1 or 2 lines following the Vlan screws me. Any ideas please?

    Read the article

  • Vim (Windows) - new help files do not work

    - by Jon
    I have downloaded several new plugins which come with a help file. None of them seem to work from within Vim. I have the Windows version of gVim and the install dir contains a folder called: vimfiles vim72 Both folder contain a plugin and doc folder. Where am I supposed to put the plugin? So when I use :help [helpfile], it says it cannot find it, but I can see it in the doc folder.

    Read the article

  • SQL Server 2005 FREETEXT() Perfomance Issue

    - by Zenon
    I have a query with about 6-7 joined tables and a FREETEXT() predicate on 6 columns of the base table in the where. Now, this query worked fine (in under 2 seconds) for the last year and practically remained unchanged (i tried old versions and the problem persists) So today, all of a sudden, the same query takes around 1-1.5 minutes. After checking the Execution Plan in SQL Server 2005, rebuilding the FULLTEXT Index of that table, reorganising the FULLTEXT index, creating the index from scratch, restarting the SQL Server Service, restarting the whole server I don't know what else to try. I temporarily switched the query to use LIKE instead until i figure this out (which takes about 6 seconds now). When I look at the query in the query performance analyser, when I compare the ´FREETEXT´query with the ´LIKE´ query, the former has 350 times as many reads (4921261 vs. 13943) and 20 times (38937 vs. 1938) the CPU usage of the latter. So it really is the ´FREETEXT´predicate that causes it to be so slow. Has anyone got any ideas on what the reason might be? Or further tests I could do?

    Read the article

  • Sphinx search: failed to unpack mysqlcompress column

    - by Un_tangable
    I have compiled sphinx 0.9.9 for debian x64. When I try to index using unpack_mysqlcompress, I get this error: "WARNING: failed to unpack column 'page_text', error=-5" The 'page_text' field is around 5Kb, and unpack_mysqlcompress_maxsize is set to 64M, yet the zlib manual says error -5 indicates Z_BUF_ERROR. Also if mysql does the decompression using UNCOMPRESS(), it works fine. Any ideas why sphinx is generating this error?

    Read the article

  • Are there Vi/Vim users who aren't touch typists?

    - by michael
    I'm trying to write a Vim tutorial and I'd like to start by dismissing a few misconceptions, as well as giving some recommendations. I don't know if I should dismiss touch-typing as a misconception, or include it as a recommended prerequisite. At the time I learned the editor, I had already been touch typing for a couple of years, so I have absolutely no idea what would be the experience of a two-fingered typist in Vim. Are you a vim two-fingered typist? what has your experience been like?

    Read the article

  • Useful Vim features

    - by Craig H
    Vim is my editor of choice, and I feel I am above average in my use of it. I do recognize, though, that the feature list of vim is huge. With this in mind, I was wondering what features you vim users out there use on a regular basis.

    Read the article

  • Fast and efficient way to read a space separated file of numbers into an array?

    - by John_Sheares
    I need a fast and efficient method to read a space separated file with numbers into an array. The files are formatted this way: 4 6 1 2 3 4 5 6 2 5 4 3 21111 101 3 5 6234 1 2 3 4 2 33434 4 5 6 The first row is the dimension of the array [rows columns]. The lines following contain the array data. The data may also be formatted without any newlines like this: 4 6 1 2 3 4 5 6 2 5 4 3 21111 101 3 5 6234 1 2 3 4 2 33434 4 5 6 I can read the first line and initialize an array with the row and column values. Then I need to fill the array with the data values. My first idea was to read the file line by line and use the split function. But the second format listed gives me pause, because the entire array data would be loaded into memory all at once. Some of these files are in the 100 of MBs. The second method would be to read the file in chunks and then parse them piece by piece. Maybe somebody else has a better a way of doing this?

    Read the article

  • which delimeter to use while spliting String

    - by London
    I need to split this line string in each line, I need to get the third word(film name) but as you see the delimeter is one big blank character in some cases its small like before the numbers at the end or its big as in front of numbers at front. I tried using string split with(" ") regex, and also \t but get the out of the bounds error. 400115305 Lionel_Atwill The_Song_of_Songs_(1933_film) 7587 400115309 Brian_Aherne A_Night_to_Remember_(1943_film) 7952 Did anyone have the same problem?

    Read the article

  • Online OCR website for processing an entire pdf file at one time?

    - by Tim
    I am looking for an online OCR website for processing a multi-page pdf file at one time. Free preferably. I know http://www.newocr.com/. If I am correct, it can only OCR one page at a time, by manually clicking "Preview" and then clicking "OCR" for each page. After each page is OCRed, I have to copy out the text result manually too. If my pdf file has 30~ pages, it will be tedious to repeat the above process for each page. I wonder if there is some other online websites that OCR a whole pdf file, without asking me for manual operation? Thanks!

    Read the article

  • How to do simple multitasked loop processing over filenames with PowerShell?

    - by Ville Koskinen
    I'm batch transcoding some 50 GB of video files on a USB hard disk which is connected to a wlan router. The drive is mapped as a network drive on my Windows 7 laptop. The speed handicap of the wlan causes some parts of the processing to become unnecessarily slow, so I would like to do the following with PowerShell: List the names of the files on the network drive to be transcoded Copy the first file to a temporary folder on my laptop Simultaneously Transcode the file in the folder Begin copying the next file from the network drive to the temporary folder After transcoding and copy have both ended, Delete the file which has been transcoded from the temporary folder Begin transcoding next file in the temporary folder Loop until all files have been processed How would I be able to do this with PowerShell? The multitasking part is an obstacle for my skill/persistence combination.

    Read the article

  • Are there ways to write php/python code to run as hooks in the Apache Request Processing pipeline?

    - by SB
    Does anybody know of any modules that provide the functionality to write python or PHP code to run as hooks in the Apache request processing pipeline? For instance, mod_perl lets me write PerlModules, which can contain handlers for the header parsing phase, content delivery, and even filters. I would like to do something similar in other scripting languages. I could write it in C, but the goal is to deploy a module that would work across a number of systems. If I deliver it as binary in C, then it would require 64/32-bit versions and some other issues. With perl, I can just require certain modules installed and mod_perl2.

    Read the article

  • What we have to measure for measuring server performance If we can't measure the server processing time from client side?

    - by AsadYarKhan
    If we can not measure the server processing time from client side then which attributes will be good to measure in client side for measuring server side performance and What attributes are important ? I know we can get the server response time, latency and Throughput etc,but how do we understand/interpret the result of server side from these attrubutes. How can we analyse that whether my code is taking lots of time,whether Web Server, whether it is because of Server Machine(H/W).how would i know that which thing needs to be upgrade or improve.Please tell me any article or any book something that I need to study or explain here If you can so I can interpret the result of server side using these attributes response time, latency and throughput.You can tell other performance attribute if I need to understand the server result.

    Read the article

  • how to fix An error occurred during the processing of a configuration file required to service this request

    - by Alex
    Just created branda new MVC 4 project and instead of expected "hello world" got following error: ==================================================== Server Error in '/' Application. Configuration Error Description: An error occurred during the processing of a configuration file required to service this request. Please review the specific error details below and modify your configuration file appropriately. Parser Error Message: Default Role Provider could not be found. Source Error: Line 244: Line 245: Line 246: Line 247: Line 248: Source File: c:\WINDOWS\Microsoft.NET\Framework\v4.0.30319\Config\machine.config Line: 246 Version Information: Microsoft .NET Framework Version:4.0.30319; ASP.NET Version:4.0.30319.272 =============================================================== Any idea how to fix this? Thanks

    Read the article

  • Does the QPI figure matter for large scale data processing using SAS?

    - by xiaodai
    At work we use SAS to manipulte large amounts of data everyday on our workstations. To give an indication of scale the largest merges we had to do was merging 24 files of 2GB in size each together into one big file (if you are familiar with SAS the files are binary compressed too!). If we upgrade our PCs to core i7 then which of Core i7 975 or Core i7 960 is better? The main difference between the two seems to be QPI. So does that affect large scale data processing such as merge datasets in SAS?

    Read the article

< Previous Page | 230 231 232 233 234 235 236 237 238 239 240 241  | Next Page >