Search Results

Search found 58486 results on 2340 pages for 'data integrator'.

Page 780/2340 | < Previous Page | 776 777 778 779 780 781 782 783 784 785 786 787  | Next Page >

  • Does a USB hub affect performance?

    - by user1018733
    I have two devices I want maximum throughput and latency with. (Midi drums and midi keyboard for example) Would connecting both to the same USB port via a hub effectively limit the maximum data transfer rate to 1/2 to each of them? I am assuming yes, but I didn't know if USB hubs had a handshaking and priority giving protocol available (e.g. let the device with the longer built up buffer of data communicate first) Thanks

    Read the article

  • Nokia PC Suite hangs at installing required drivers and doesn't connect to phone

    - by SpongeBob SquarePants
    I am unable to connect my Nokia 2690 to my PC via data cable using Nokia PC Suite under Windows XP SP3 Pro 32 bit. Every time I connect my mobile using the data cable, the Nokia App shows a window where it says it's "Installing the required drivers. Please wait". But even after several hours my phone never conects to my PC and the screen remains there idle. I checked in he task manager, Nokia PC Suite didn't go unresponsive.

    Read the article

  • Few question on windows explorer properties(win 7)

    - by Nrew
    I've red this article from howtogeek, but it didn't explain this one which is placed in the target portion when you right click on windows explorer and click properties: %windir%\explorer.exe shell:desktop\Inbox And why does local disk E: shows up when I have this one: %windir%\explorer.exe shell:E:\FINAL SAVE DATA I don't really get the code, especially the part in shell: desktop\Inbox. What's that supposed to mean. How do I change it so that when I click on the Windows Explorer shortcut, I get to see this location: E:\FINAL SAVE DATA

    Read the article

  • Does SDHC have any write (ECC) error recovery ?

    - by marc
    What happen if SDHC card get write error (damaged cell / bad sector) ? Whole card is unusable (to trash, all data written to that sector now and in future will be lost) ? or rewrite sector (flash memory get corrupted when writing so maybe have any function to check if sector was written successfully) to another and mark as fault as unusable what will be seen as reduction of capacity but no data lost. I have to do some research about SD card-s on disk less machines. regards

    Read the article

  • Piping Objects From Format-Table to Export-CSV Has Unexpected Results

    - by makerofthings7
    I'm running this command to get a list of all Activesync users and export them to C:\activesync.csv Get-ActiveSyncDevice | Get-ActiveSyncDeviceStatistics | sort-object status, devicetype , lastsyncattempttime | ft FirstSyncTime ,LastPolicyUpdateTime ,LastSyncAttemptTime ,LastSuccessSync , DeviceType , DeviceID, DeviceAccessState, Identity -a | Export-Csv c:\activesync.csv The problem is that the CSV data doesn't match the console display if I omit the trailing | c:\activesync.csv ... where the data and columns displayed don't match. Is this a bug in powershell?

    Read the article

  • HP E200i Controlller RAID Configuration fon Win2008 Ent, Sql Server, IIS Apps etc need opinion

    - by mn
    Hello, Actually I run RAID 5 (4 x SAS drives) with Win 2008 Ent(1x host) Win 2008 End(3x guest) Sql Server 2005 Std (on guest) 3 x asp.net applications (on guest) I bought 3 x drives to create additional array (on same controller E200i, I am waiting now for confirmation is it possible to have 3 raids in same controller) I am planning to have 2 x RAID5 (if it is possible) first RAID 5 with all vhd files, systems etc second RAID 5 all data files and transaction logs I am looking for opinion how to optimize data layer (seven drives, one controller).

    Read the article

  • Excel 2007 save import steps on csv file?

    - by Chris Marisic
    I have a csv file that constantly needs opened into Excel and then have the data copied over to a separate workbook. I find the process of having to click through all of the dialogs, setting the text identifier, setting the columns to all be text extremely tedious. In many actions with data like this in regards to MSSQL or Access the program will ask you if you wish to save these steps however Excel doesn't readily ask that. Is there any way to get a comparable usage with Excel?

    Read the article

  • Stop excel from converting copy-pasted number/text values to date

    - by Tomas
    I'm copy-pasting some data from html table into excel. But excel automatically converts some text or number values to date! When I change the format, the number is perversed, the number is something like 4112523 (excel probably interprets the cell as date and then converts to number or something like that...) There is a trick for importing CSV files, but is there any solution when you are pasting your data directly from a web browser?

    Read the article

  • Lots of dropped packages when tcpdumping on busy interface

    - by Frands Hansen
    My challenge I need to do tcpdumping of a lot of data - actually from 2 interfaces left in promiscuous mode that are able to see a lot of traffic. To sum it up Log all traffic in promiscuous mode from 2 interfaces Those interfaces are not assigned an IP address pcap files must be rotated per ~1G When 10 TB of files are stored, start truncating the oldest What I currently do Right now I use tcpdump like this: tcpdump -n -C 1000 -z /data/compress.sh -i any -w /data/livedump/capture.pcap $FILTER The $FILTER contains src/dst filters so that I can use -i any. The reason for this is, that I have two interfaces and I would like to run the dump in a single thread rather than two. compress.sh takes care of assigning tar to another CPU core, compress the data, give it a reasonable filename and move it to an archive location. I cannot specify two interfaces, thus I have chosen to use filters and dump from any interface. Right now, I do not do any housekeeping, but I plan on monitoring disk and when I have 100G left I will start wiping the oldest files - this should be fine. And now; my problem I see dropped packets. This is from a dump that has been running for a few hours and collected roughly 250 gigs of pcap files: 430083369 packets captured 430115470 packets received by filter 32057 packets dropped by kernel <-- This is my concern How can I avoid so many packets being dropped? These things I did already try or look at Changed the value of /proc/sys/net/core/rmem_max and /proc/sys/net/core/rmem_default which did indeed help - actually it took care of just around half of the dropped packets. I have also looked at gulp - the problem with gulp is, that it does not support multiple interfaces in one process and it gets angry if the interface does not have an IP address. Unfortunately, that is a deal breaker in my case. Next problem is, that when the traffic flows though a pipe, I cannot get the automatic rotation going. Getting one huge 10 TB file is not very efficient and I don't have a machine with 10TB+ RAM that I can run wireshark on, so that's out. Do you have any suggestions? Maybe even a better way of doing my traffic dump altogether.

    Read the article

  • Printing barcode labels

    - by BoundforPNG
    I am trying to print barcodes for a library. I have generated a list of sequential numbers and copied them into Word 2003. I want Library name, barcode font, barcode number printed in an Avery 5160 label sheet. I can get it to work by going to tools, letters and mailing, Envelopes and Labels, labels tab and type in what data I want. I don't see how I can pull this data from the list in the word document rather than just what I type.

    Read the article

  • NFS host is not exporting the "share"

    - by user1345260
    I have a NFS Server: usanfsd01 And a remote machine: usafssd01 I tried mounting a directory from usafssd01 onto usanfsd01 by adding the following line to /etc/fstab as root usanfsdo1:/home/dblogs /home/data/dblogs nfs rw 0 0 And when I run the following command to see if NFS is exporting the share, it's not shown showmount -e usanfsdo1 Can someone please help? Also, a point of interest would be there is another mount that works on the same servers and thats defined as below in the /etc/fstab usanfsdo1:/home/files /home/data/files nfs rw 0 0 /etc/exports /nfs/home/dblogs 'IPADDRESS'(rw,no_root_squash)

    Read the article

  • error while loading shared libraries Dicom Store SCU / Echo SCU

    - by David Just
    I am running a dicom receiver on a Centos 6 box on top of a Xen server. If I attempt to send data to it from a remote server I get the following error: storescp: relocation error: /lib/libresolv.so.2: symbol memcpy, version GLIBC_2.0 not defined in file libc.so.6 with link time reference If I send data to the server locally it works, but sending to it from remote gives the above error. I do not think this is a problem that is specific to the storescp server.

    Read the article

  • SSL totally stopped working in Windows

    - by Dims
    Apparently, on my notebook, I have suddenly lost any ability to use network connections, involving SSL and/or data encryption, provided my MS: 1) remote desktop connections: Because of an error in data encryption, this session will end 2) browse HTTPS sites: Can't browse HTTPS pages. TLS error 3) communicate over WiFi, while wired is ok Is there any possible one central reason for all of these problems in Windows? Third party applications, like Putty, works fine. Is it possible to reset/repair certificate store or something in Windows?

    Read the article

  • Event log message size 31885? Windows 2008

    - by testuser
    We recently upgraded our production boxes to Windows 2008 from Windows 2003 servers. Everything works fine except the event logging. We log at max 32000 bytes of data for each message On 2008 servers, event logging fails if number of characters is greater than 31885. Is this new limit on Windows 2008 R2 servers? Any help appreciated. On Win 2003 servers, I am able to log 32000 bytes of data for each log entry.

    Read the article

  • How to unmangle PDF format into a usable text or spreadsheet document?

    - by Chuck
    Upon requesting some daily/hourly sales data from a coworker who is responsible for such requests, I was given a series of PDF files. The point of sale program that is used, for some reason, answers requests for this type of information in the form of PDF files. The issue: The PDF files look to be in a format that should easily be copy and pasted into a spreadsheet. There are three columns that look to be neatly organized across two pages. When copy/pasting the first page, all three columns from the PDF's first page are dumped into a single column consisting of the Date followed by the Hours for the transactions on that day. The end of this Date/Time information is followed by all of the Total Sales values that should be attached a Date and Time of the transaction. (NOTE: There are no duplicated Dates in the Date column, ie, Multiple transactions for a day only have one yyyy/mm/dd listed for the first row but not the following rows.) While it was a huge pain, it was possible to, in about four or five steps, get the single column of data broken out into three columns that matched the PDF. The second page of the PDF file, when attempting to copy/paste into a spreadsheet, creates a single column with the first third of the cells being the Dates from the PDF, the second third of the cells being the Hours of the transactions and the final third of the cells being filled with the Total Sales. After the copy/paste there is no way to figure out which Hours belong to which Dates or Total Sales due to the lack of the duplicated Dates in the Date column as mentioned above. My PDF-fu is next to non-existent. I've just now started to work with PDF editors and some www.convertmyPDFforfree.com websites, so far, with absolutely nothing remotely coming anywhere near usable output. (Both methods have so far done nothing but product blank documents.) Before I go back and pester my co-worker into figuring out a way to create a report in some other format than PDF, is there any method by which to take the data that looks to be formatted correctly in a PDF and copy/paste it into a spreadsheet that will look the same? I appreciate any help that can be made available. The sales data isn't so sensitive that I couldn't part with a bit to let somebody actually see what it is that needs to be dealt with, just let me know. The PDF's are less than 100kb each so sending them shouldn't be a burden to any interested party.

    Read the article

  • How could I determine which SMB client/session has a specific file open on a Server 2008R2 Windows file server?

    - by Rasmir
    What I need a way to associate a client name or IP address with an open file, so that I can cleanly close the file for maintenance. NET SESSION doesn't show the names of open files and NET FILE doesn't show the client which has the file open. I had hoped that I could cross-reference the data from these two commands, but that doesn't seem doable. Everything else I've see provides the same data as these commands, with no apparent way to determine which client machine has the file open.

    Read the article

  • Server with 3 Disk, what's the best HD Configuration?

    - by aleroot
    I Have an HP Server with a quad core Opteron and 3 Disk 250Gb S-ATA Disk, i'm thinking about what's the best configuration of the disk for performance and reliability. There is mainly 2 scenario : -RAID 5 with these 3 HD (on the the array 100GB Partition for OS, Other Space for Data Partition) -RAID 1 + 1 Disk for OS (one single Disk OS Installation, RAID 1 Array for a Data Partition) What's the best configuration ? In the Server Run MySQL and Small Document File server, the OS to be installed is Windows Server 2008 ...

    Read the article

  • Realtime MS SQL Database transfer

    - by user34402
    Hello, I have MS SQL 2005 Standard that is used actively (new data is added every minute). I need to transfer it to another server (new hosting provider), without losing the data and without downtime. Is there any way to do it?

    Read the article

  • NFS server Windows 2008 - mounting via linux - input/output error help.

    - by pablo
    I want to try mounting a folder with NFS - I set up the NFS sharing on a windows 2008 R2 server, specified hosts in the NFS permissions (by IP address) and mounted via /etc/fstab it mounts, but when I try to list the folder, I get 'input/output error' the owner/group on the local mount point look weird too? drwx------ 2 4294967294 4294967294 4096 2011-02-10 19:15 data/ I mounted in /etc/fstab as: 10.0.6.55:/share$ /media/data nfs soft,intr,rsize=8192,wsize=8192 What am I doing wrong?

    Read the article

  • Is it perfectly safe to install grub bootloader on regular partition?

    - by Flint
    One of the methods to do dual booting Windows with Linux OS is by installing grub boot loader onto Linux partition so you can retain Windows boot loader and let Windows handles the dual booting process. What's the odd that grub bootloader could partially overwrite the data at the beginning of the Linux partition and corrupt the file? Does grub actually check if there's a data at the beginning of the partition and move it to other location on the partition before writing its bootloader?

    Read the article

  • RADIUS PROTOCOL

    - by Vijay.J055
    Hi, I am using radius protocol to for sending some values from client to server. in that i am using vendor specific value pairs,and definng our own types. but the value length for Vendor-speciifc data is 255,but our data lenght is crossing it. please can any one tell me how to incorporate the dta with length more than 255 bytes thanks in advance

    Read the article

  • how to return my losed drive

    - by sama
    the drive D in my computer has been losd i opend my computer but i didnot find it i don't know wher it is and i don't know how to return it i go to disck manaement i found it as a free space then i try to make it NTFS but it was needed to format it but i don't want to format it i need the data can any one help me to return my drive without losing my data

    Read the article

  • Can I mix yum and apt-get safely?

    - by leonbloy
    I understand that yum and apt-get operate on top of rpm, so that the data about installed packages in a linux system is responsability of rpm; so that neither yum nor apt-get keep their own data about installed packages . Is this true ? It is safe to install some package using yum and install another (perhaps related) package using apt-get (or viceversa)?

    Read the article

< Previous Page | 776 777 778 779 780 781 782 783 784 785 786 787  | Next Page >