Search Results

Search found 19525 results on 781 pages for 'say'.

Page 379/781 | < Previous Page | 375 376 377 378 379 380 381 382 383 384 385 386  | Next Page >

  • Man pages in Linux

    - by Ayos
    I don't seem to have all the man pages that I need. For example, my college computers (running Fedora 14) have man pages for ASCII, all the standard C libraries (stdlib.h, stdio.h) and so on and so forth. I wish to "install" these pages, after looking up on the Internet I couldn't really find anything that made sense. How can I get, say, the man-page of ASCII (I know it's not really a command or a daemon or anything like that, but typing man ASCII on the college computer gets me a page with the ASCII value table + a little more information). I don't want to keep using the Internet for looking up man pages every time I need to look up a function, function prototype or the ASCII table or something like that.

    Read the article

  • rsync generates very much traffic

    - by user109459
    I use rsync for backing up one of my servers with 4GB of files. When I now try to transfer these files the traffic for the files isn't the estimated 4GB. It is a lot higher. It's about 60GB. I also checked the traffic on my server, backup server and router and all three say that there was a traffic of 60GB. But at the end rsync says that it only has transfered 4GB. Another problem is that I can't debugg it because the problem occures randomly.

    Read the article

  • How do I pull a backup from a Linux server to my Windows PC using rsync?

    - by Nogwater
    I'm currently using sftp to download nightly backups (.tar.gz) from my web host to my desktop computer. I think I'd like to switch to rsync to minimize the bandwidth (and time). I have cygwin installed on my PC, but don't use it for much. I have shell access to my web host via ssh (PuTTY). Let's say my source directory is myserver.com:/home/username/backups/, I want to grab all of the .tar.gz files from there, and I want to save them to C:\Backups\ locally.

    Read the article

  • Best method to implement a filtered search

    - by j0N45
    I would like to ask you, your opinion when it comes to implement a filtered search form. Let's imagine the following case: 1 Big table with lots of columns It might be important to say that this SQL Server You need to implement a form to search data in this table, and in this form you'll have several check boxes that allow you to costumize this search. Now my question here is which one of the following should be the best way to implement the search? Create a stored procedure with a query inside. This stored procedure will check if the parameters are given by the application and in the case they are not given a wildcard will be putted in the query. Create a dynamic query, that is built accordingly to what is given by the application. I am asking this because I know that SQL Server creates an execution plan when the stored procedure is created, in order to optimize its performance, however by creating a dynamic query inside of the stored procedure will we sacrifice the optimization gained by the execution plan? Please tell me what would be the best approach in your oppinion.

    Read the article

  • Setting primary IP for VPS (out of 2 available)

    - by Flyer
    I've got a little problem. My VPS (Debian box) has 1 primary IP address, but now my hosting provider decided to assign me another one due to problems with first one. Let's say new IP is 91.222.152.20 and old one 91.221.190.170. How should I make 91.222.152.20 primary one, so that everything works on this address instead of 91.221.190.170 (like ftp, mysql etc) I tried editing /etc/newtwork/interfaces setting iface eth1 inet static address 91.222.152.20 netmask 255.255.252.0 network 91.222.152.0 broadcast 91.222.155.255 gateway 91.222.152.1 I know that network is 91.222.152.0/22 so I assume that above configuration should be right. Saved, rebooted and that ended pretty bad. Couldn't even ping box on old or new IP. Had to revert to working one. So question is. What am I doing wrong and how to do it right?

    Read the article

  • Hooking domain to home server (WinServer2008) with specific port

    - by user1071461
    Alright, I'm asking two things here. First of all, if i purchase a domain let's say myhomeserver.com, am I able to make the default port go through a different port instead of the default port 80? (that is without having to do myhomeserver.com:5000 for example). Also this should be without blocking other ports (so no stealth forwarding to myhomeserver.com:5000 i think) Secondly, How could I go about hooking a domain to a windows 2008 server? I've seen it on linux but no clue how to do it on windows if it's even possible. I know I'm asking a lot here, just some tips are appereciated. Also, yes I know, using a home server is horrible for security and preformance and whatnot, I understand this already, thanks ^^

    Read the article

  • Route using certain IP address

    - by spa
    I have a server with two public IPs. Both IPs are added to eth0 using ip addr add. Now I'd like to contact a server which uses IP address filtering. Only requests are allowed which use the second IP address. Is there are way to set this up using the standard route command in Linux? I guess that's not the case. So the only solution I see right now: Setup a virtual device let's say eth0:0 and bind the second IP address to it. Then I can reference the device in the route command. Edit: I can't use the second IP as primary one easily as this IP is used as failover IP.

    Read the article

  • Work @ Java shop. Tasked to redo intranet. I only know PHP - CTO says use that; Ops Says no - we have JSP, use that - Thoughts?

    - by Mackysback
    So I work as a project manager and was given the opportunity to redo the intranet and Internet site... I'd love to , great addition to my resume. However.... I am not familiar with java nor JSP... I am fairly proficient with PHP and MySQL ... Also my intention was to use a CMS that I have experience with, wordpress, drupal or joomla... The site would be very simple so I was thinking wordpress would be fine for this. I told management that I would need it to run on PHP and they gave me their blessing... Now the disconnect bw it and management ... I spoke with ops and they told me that we have java and JSP for that purpose although the IT guy did say "oh but I do know that php and mysql are gaining popularity" That said ... I do not understand much as far as systems architecture... We do self host and run websphere with IIS and jboss with tomcat. Any advice or suggestions? Thanks ! Is my request that unfeasible?

    Read the article

  • Using git with cgit for decentralized/centralized development

    - by polemon
    I plan to use git for hosting my projects on my server. I've read about cgit, git-daemon, and I more or less decided to use those tools. But general use is still kind of confusing for me. What do I need to set up on the server, to push my files onto it. And when the files on the server are newer as the files on my computer, how do I merge them? Also, I use, say, two computers where I develop. How do I merge from one computer to the other? Also, when two people are working on the same project, how do they merge their local repos from one another? As you probably can tell by now, I come from SVN, but I've worked with Mercurial and now I'd like to test git.

    Read the article

  • Experience with Intel X25-M 160GB and Oracle

    - by derobert
    We're considering building an Oracle database with 12 Intel X25-M G2 160GB drives in software RAID10. It'd be running Linux. Database gets some very heavy write activity during the early morning data load, other than that it is mostly read-only (and the read load is fairly minimal). We're currently running on 11 150GB Velociraptors (also Linux software RAID10), and are hoping the X25-M will speed up the data load. We currently have redo on different disks than the rest of the data. I'm wondering a few things: Any experience with using X25-M drives for databases? The X25-E are unfortunately beyond our budget. Would it hurt to separate redo off to some magnetic (non-SSD) drives, say 2 (raid1) or 4 (raid10) Seagate Constellations?

    Read the article

  • Is possible to arbitrarily register names to the same public IP?

    - by Alex. S.
    I registered a domain, lets say mysite.com (for example), then, results that somebody else has an A record from anotheraddress.com pointing to the same IP address of mine (in a VPS in linode.com) What can I do to avoid this???, I mean, I would prefer reject accesses from anotheraddress.com to my site. I just know only by casualty putting my genuine domain name on this http://www.domaintools.com/reverse-ip/ My DNS server is name.com, and the DNS server pointing to the my public IP is from GoDaddy. Is possible to register arbitrarily names to the same public IP? Can I use my DNS record with mysite.com to point to 209.85.133.147 (google.com), for example?

    Read the article

  • How can I force certain applications to use specific network connections?

    - by snicker
    Let's say I have two active network connections that let me out to the internet. I want certain applications to only use Network Connection 1, while some others should use Network Connection 2. Is this possible in Windows XP? If so, how can it be done? The main reason I wish to do this is I want to use a tethered phone's network for certain applications and an ethernet connection for others. Certain ports and networks are blocked by the ethernet connection, whereas they are not on my tethered phones connection.

    Read the article

  • ripping interlaced video to 60fps (frames per second)

    - by cloneman
    It would appear that Broadcast Television is often 1080i30 (60 fields/s), and non-movie DVDs (example, instructional or TV shows) are 60fields/s as well, at some lower resolution (480i?) However, almost all video that ends up on the internet, whether in x264-encoded content, Youtube, etc. is 30 frames per second, that is to say , it is progressive scan. However, when you watch content on your TV, I'm guessing the TV converts it to progressive for you, but the end result is a very fluid picture that "feels" quite a bit like 60frames/s. What is the best way to obtain this result when ripping interlaced content sourced from TV or DVDs? Can I rip a DVD that is 60 fields per second to 60 frames per second? I would imagine classic deinterlace filters do not do this, they merge fields and create a 30fps output. I'm using handbrake.

    Read the article

  • Interfaces: profit of using

    - by Zapadlo
    First of all, my ubiquitous language is PHP, and I'm thinking about learning Java. So let me split my question on two closely related parts. Here goes the first part. Say I have a domain-model class. It has some getters, setters, some query methods etc. And one day I want to have a possibility to compare them. So it looks like: class MyEntity extends AbstractEntity { public function getId() { // get id property } public function setId($id) { // set id property } // plenty of other methods that set or retrieve data public function compareTo(MyEntity $anotherEntity) { // some compare logic } } If it would have been Java, I should have implemented a Comparable interface. But why? Polymorphism? Readbility? Or something else? And if it was PHP -- should I create Comparable interface for myself? So here goes the second part. My colleague told me that it is a rule of thumb in Java to create an interface for every behavioral aspect of the class. For example, if I wanted to present this object as a string, I should state this behaviour by something like implements Stringable, where in case of PHP Stringable would look like: interface Stringable { public function __toString(); } Is that really a rule of thumb? What benefits are gained with this approach? And does it worth it in PHP? And in Java?

    Read the article

  • PcLinuxOs demands I use only one repository at time. Is it right?

    - by m33600
    I come to your presence with this question that is paralyzing my coding efforts. PclinuxOs was my distro of choice for reliability, but it is jealous and does not permit me to add repos from, say, Debian. The wiki is clear advising on using just one repo, and I end up not finding what I used to find on normal Debians. Multimon, the audio decoder, for example (my other question) is not there. When I try to install multimon with hammer and plies, it returns errors of all kinds. Is there a way to safely and temporarily add a repository, make the install and remove the repo, returning pclinuxos to its stable state?

    Read the article

  • nginx serves broken characters

    - by Andrew123321
    I have nginx 1.2.0-1 on debian 6.0.5. I have an empty file (let's say test.css). I add A and newline, works fine, B and newline, works fine,... D, works fine. I add E and I got three broken characters ("???"). When the file has got more content nginx would usually "cache" it and then output the cached file minus some characters plus some of these broken characters. Do you have any idea how to deal with this? Thank you

    Read the article

  • How to export or view audio file references in a PDF?

    - by redshift
    I have a an interactive PDF file that is over 90+ pages long. Each page is a map with city names that contains a Spanish pronunciation of that city in a .wav file. I'd say there are about 10-15 audio files for each map which comes out to 1000+ audio files. Is there a way to extract/export a list of the sound file names associated with each map? I tried to save the PDF to an HTML file, but it only exported images and text, and because the audio files were embedded in the PDF, the file names did not carry over to the HTML file. Any other ideas? I need to see what audio file goes with what map/page.

    Read the article

  • Software Architecture: How to divide work to a network of computers?

    - by Morpork
    Imagine a scenario as follows: Lets say you have a central computer which generates a lot of data. This data must go through some processing, which unfortunately takes longer than to generate. In order for the processing to catch up with real time, we plug in more slave computers. Further, we must take into account the possibility of slaves dropping out of the network mid-job as well as additional slaves being added. The central computer should ensure that all jobs are finished to its satisfaction, and that jobs dropped by a slave are retasked to another. The main question is: What approach should I use to achieve this? But perhaps the following would help me arrive at an answer: Is there a name or design pattern to what I am trying to do? What domain of knowledge do I need to achieve the goal of getting these computers to talk to each other? (eg. will a database, which I have some knowledge of, be enough or will this involve sockets, which I have yet to have knowledge of?) Are there any examples of such a system? The main question is a bit general so it would be good to have a starting point/reference point. Note I am assuming constraints of c++ and windows so solutions pointing in that direction would be appreciated.

    Read the article

  • How come there is still so much programming work?

    - by jd_505
    First I'd like to say that I am not pretty sure that this question will meet the guidelines. I think it can goes under the "Freelancing and business concerns" bullet, but I am not sure. Anyway, I will give it a shot. I wonder how the programming jobs hasn't yet "dried" because of the software evolution, for example, I am a developer myself, which means that I do care about software (I mean I am not of the type of guys that needs a computer mainly to just browse the Internet), and still I wouldn't mind if I never receive any more updates on my Ubuntu machine. I find that it provides everything I need, and while the updates provide various bug fixes/improvements, I wouldn't mind using it with it's current state for the rest of my life, for 2 years of Ubuntu usage I have never bumped at a serious bug/problem. Another example is Windows, almost half of it's users still use XP, which is practically ancient, yet they find it satisfying all their needs (and I agreee with them). I could go with many more examples, but by now you are understanding my point and my question. While new "trends" appears all of the time (like a new mobile OS) which runs on new platforms and requires some fresh development work, still the majority of the software effort goes in to what I consider as "completed projects", or at least a state of a project which is enough to be considered as completed. Do you have an explanation ? I can't think of the right tags for this question, please edit it the way you find it to be most appropriate. Thanks.

    Read the article

  • How do i keep a newly started program from taking focus?

    - by Jugglingnutcase
    Say i'm coding in emacs and want to start up a music program. Because it takes too long to start up i go back to coding and type away. When the music application starts up, the focus is stolen (gasp! stolen!) away from emacs and goes to the music application, often mid-thought. Is there any way to keep this from happening and have the newly started application not have focus until i see that it's up and ready to be used? Besides getting rid of my ADD of course. Or getting an impossibly fast computer that can keep up with my mind. i'm using a Windows XP system, but i will soon have a Windows 7 system, and i have Linux at home.

    Read the article

  • How to backup old emails locally in Thunderbird and then remove them from IMAP server?

    - by saicode
    I am using Godaddy IMAP email with Thunderbird as my desktop client on Windows 7. The email service has unlimited mailbox size but the local Thunderbird is having troubles due to the large size of the inbox/outbox. I would like to take out old emails from IMAP server and backup them locally. After backing up the old email I would like to delete old email (older than let's say, 2012) from the server. Also I'd like to have them accessible from the local backup if ever needed in the future. This way I might be able to make Thunderbird fast and problem free. Problem is, I am not able to find any instruction to do this in an automated way based on dates etc. I can find some links for Archiving, Compacting and Backup. But unable to find any tutorial about how to backup and archive it locally and delete the original emails from the server.

    Read the article

  • How to fix legacy code that uses <string.h> unsafely?

    - by Snowbody
    We've got a bunch of legacy code, written in straight C (some of which is K&R!), which for many, many years has been compiled using Visual C 6.0 (circa 1998) on an XP machine. We realize this is unsustainable, and we're trying to move it to a modern compiler. Political issues have said that the most recent compiler allowed is VC++ 2005. When compiling the project, there are many warnings about the unsafe string manipulation functions used (sprintf(), strcpy(), etc). Reviewing some of these places shows that the code is indeed unsafe; it does not check for buffer overflows. The compiler warning recommends that we move to using sprintf_s(), strcpy_s(), etc. However, these are Microsoft-created (and proprietary) functions and aren't available on (say) gcc (although we're primarily a Windows shop we do have some clients on various flavors of *NIX) How ought we to proceed? I don't want to roll our own string libraries. I only want to go over the code once. I'd rather not switch to C++ if we can help it.

    Read the article

  • How to maximize parallel download from S3

    - by StCee
    I got a lot of images to load from Amazon S3 on a single page, and sometimes it takes quite some time to load all the images. I heard that splitting the images to load from different sub-domains would help parallel downloads, however what is the actual implementation on that? While it is easy to split for sub-domains like static,image, etc; Should I make like 10 sub-domains (image1, image2...) to load say 100 images? Or is there some clever ways to do? (By the way I am considering using memcache to cache the S3images; I am not sure if it is possible. I would be grateful for any further comments. Thanks a lot!

    Read the article

  • Does Google Chrome officially work on Windows 7 64-Bit Yet?

    - by Nick Josevski
    As soon as I jumped onto one of the beta releases for Windows 7, I tried to install Google Chrome. Being on a 64-bit installation it came up with a 'non-supported OS' or some error (can't remember). Having a look around at the time I saw lots of posts/tips about just appending --in-process-plugins to the shortcut for chrome, after trying this and still not having luck, I found more posts including what seemed ones from the Chrome developers saying this was not wise and exposed a security risk. So does anyone have a well sourced answer, as to what's holding up Win 7 64-bit support in Chrome, or better yet an "official" answer to say that it is supported in Win7 x64 RTM and works well now...

    Read the article

  • Does /boot safe on top of a lvm LV (logical volume)?

    - by fantoman
    Title already asked the question. More specifically, I read in some documents that logical volumes are nice in general but not for /boot in a linux system. They say that bootloaders don't understand LVM volumes, so create a separate partition for /boot out of lvm. I recently installed Ubuntu server (9.10) for my home server, but by default /boot is created in the LVM. Everything is fine now, but I am not sure it is safe to use /boot in LVM. Second question is do I really need a physical partition (volume)(pv) for /boot or is it equally fine if I put it into a logical volume (lv) on top of a single shared volume group. Thanks in advance.

    Read the article

< Previous Page | 375 376 377 378 379 380 381 382 383 384 385 386  | Next Page >