Search Results

Search found 27819 results on 1113 pages for 'linux intel'.

Page 291/1113 | < Previous Page | 287 288 289 290 291 292 293 294 295 296 297 298  | Next Page >

  • Browser-based Operating System

    - by Ross Peoples
    I have a bunch of touchscreen machines that I want to display a webpage on and have users interact with the webpage via the touchscreen. Right now, this is done with a full-blown OS with a browser set to run at startup. I think maybe the ideal solution is to use a Linux-based OS that boots up, starts X, then starts a web browser (Chrome, Firefox, or whatever) in full screen mode. What kind of options do I have? I really want to avoid using a full-blown OS like I do now. It looks unprofessional and takes a while to boot up. I was thinking maybe Chrome OS or something, but I wouldn't know how to set it up for my purposes, since it's still designed to be used as a desktop OS instead of a kiosk-type OS.

    Read the article

  • fsck: FILE SYSTEM WAS MODIFIED after each check with -c, why?

    - by Chris
    I use a script to partition and format CF cards (connected with a USB card writer) in an automated way. After the main process I check the card again with fsck. To check bad blocks I also tried the '-c' switch, but I always get a return value != 0 and the message "FILE SYSTEM WAS MODIFIED" (see below). I get the same result when checking the very same drive several times... Does anyone know why a) the file system is modified at all and b) why this seems to happen every time I check and not only in case of an error (like bad blocks)? Here's the output: linux-box# fsck.ext3 -c /dev/sdx1 e2fsck 1.40.2 (12-Jul-2007) Checking for bad blocks (read-only test): done Pass 1: Checking inodes, blocks, and sizes Pass 2: Checking directory structure Pass 3: Checking directory connectivity Pass 4: Checking reference counts Pass 5: Checking group summary information Volume (/dev/sdx1): ***** FILE SYSTEM WAS MODIFIED ***** Volume (/dev/sdx1): 5132/245760 files (1.2% non-contiguous), 178910/1959896 blocks

    Read the article

  • My datacenter is unable to add PTR/rDNS record, then how can I prevent the mail outgoing from going to spam folder?

    - by gilzero
    I am having problem that mail sent out from my server all goes to recipient's spam folder. I am running Drupal sites on Linux server. CentOS w/ cPanel. Our users cannot receive email as the mail went to the spam folder. (such as registration email, contact form email) I was advised that I need to have PTR/rDNS record added for my host. I then contact my datacenter to add PTR/rDNS thing, unfortunately, the datacenter said they are unable to do it. So what can I do? Any other ways I can fix the problem? Thank you!

    Read the article

  • Best practice for scaling a single application source to multiple nodes

    - by Andrew Waters
    I have an application which needs to scale horizontally to cover web and service nodes (at the moment they're all on one) but interact with the same set of databases and source files (both application code and custom assets). Database is no problem, it's handled already with replication in MongoDB. Also, the configuration of the servers are the same (100% linux). This question is literally about sharing a filesystem between machines so that its content is always correct, regardless of the node accessing it. My two thoughts have so far been NFS and SAN - SAN being prohibitively expensive and NFS seeing some performance issues on the second node with regards to glob()ing in PHP. Does anyone have recommended strategies or other techniques that don't involved sharding data across nodes or any potential gotchas in NFS that may cause slow disk seek times? To give you an idea of the scale, the main node initialises it's application modules in ~ 0.01 seconds. The secondary is taking ~2.2 seconds. They're VM's inside a local virtual network in ESXi and ping time between them is ~0.3ms

    Read the article

  • Remote Control Home PC from Corporate Work PC

    - by muncherelli
    Here is my situation: I am currently on a Windows XP workstation at work. I have an android tablet that I use to splashtop into my home PC. I would like to be able to use my work keyboard and mouse to control my home PC while I am splashtop'd into it using my tablet. My work PC is on a corporate LAN, and not on the same network as my tablet. The company I work for provides wifi for personal devices, but they are not accessable to the internal network. I thought about going the Synergy route, however that would require my home PC to be able to connect to my work PC which isn't really possible. The opposite would work though, if I could reverse connect the server to the client, but the Synergy software doesn't really support that. I do have a couple linux boxes running at home, so I can ssh into my home network and tunnel ports via SSH if needed. With what I have, how can I accomplish seamless keyboard and mouse sharing between my work PC and either my home PC or my android tablet?

    Read the article

  • How do I configure networking on CentOS 6 running on hyper-v?

    - by LonelyLonelyNetworkN00b
    I'm not using legacy adapters, and i've installed Linux Integration Components 3.2. THe problem i'm facing is that the command 'setup' or 'system-config-network' doesn't list any network interfaces. If i run ifconfig -a i can see both the network cards i've attached. By setting a ip using ifconfig i can get network connectivity. The problem is that it's not persistent after a reboot. I'm a 100% centos newbie, but I figure it has something to do with that the centos installer couldn't see the NICs on install. How can I fix this?

    Read the article

  • Where can I configure wireless to be passed on to the Virtualbox guest?

    - by huahsin68
    I have WinXP install in virtualbox which host in Linux. I have a TP-Link (TP-WN321G) USB wifi adapter and have the driver installed inside WinXP. When I plug-in the wifi adapter, there is an option show "Ralink 802.11g WLAN [0101]" in the virtualbox's USB icon, tick on that option, the Device Manager able to detect the hardware which shows TP-Link, but when look into the properties, it says there is no driver was install. I did try to install Ralink driver but still no luck. Just curious why my wifi adapter is TP-Link, but the option show Ralink? May I know how can I emulate the wireless network inside WindowsXP?

    Read the article

  • Is it possible to modify a color scheme and windows decorations in Xfce4?

    - by Juhele
    just testing PCLinuxOS Phoenix XFCE Edition 2011-07 which I would like to install on my grandpa's PC instead older PCLOS 2009 with KDE3 (which is almost impossible to upgrade). The PC is relatively older one (Sempron 2200+ CPU, MSI K7N2GM2 board with integrated GeForce 440MX series graphics, 1GB RAM and 80GB IDE HDD) and I thought that it is too weak for KDE4. I already used Xfce in the past (Sam Linux 2006 and 2007), but in new Xfce4 I am not able to somehow change the windows color scheme and the windows decorations - the settings manager offers me switching between preinstalled themes but it is possible to modify them with some GUI?

    Read the article

  • Conference Calling Meetme.conf PROBLEM - Astarisk

    - by Zafer
    I am facing problem in conference calling (OS ubuntu latest) ... I created a conference Room and an extension no .. when i dial from my softphone (3cx or X-Lite) it says "YOU ARE ENTERING CONFERENCE NO 1234. YOU ARE CURRENTLY THE ONLY PERSON IN THE CONFERENCE) and then a music for a second and then shows connected... Now when i dial form another soft phone the same extension it goes to second conference room and plays the same message .. if i create only one conference room .. the second no. didn't connect to it and says no conference room. CAN some one help me out .... I just want to create a conference room OR extension where 4-5 or 6 people can call and make a conference ... i am not very much expert in linux and asterisk so simple steps to achieve my goal .. Thanks

    Read the article

  • May the file size returned by stat be compromised?

    - by codeholic
    I want to make sure that nobody changed a file. In order to accomplish that, I want not only to check MD5 sum of the file, but also check its size, since as far as I understand this additional simple check can sophisticate falsification by several digits. May I trust the size that stat returns? I don't mean if changes were made to stat itself. I don't go that deep. But, for instance, may one compromise the file size that stat returns by hacking the directory file? Or by similar means, that do not require superuser privileges? It's Linux.

    Read the article

  • Watch Filesystem in Real Time on OS X and Ubuntu

    - by Adrian Schneider
    I'm looking for a CLI tool which will watch a directory and spit out the names of files that change in real time. some_watch_command /path/to/some/folder | xargs some_callback I'm aware of inotify (inotify-tools?) and it seems to be what I need, but I need something that is both Linux (in my case Ubuntu) and OSX compatible. It doesn't need to be lightning fast, but it does need to trigger upon changes (within a second is reasonable). Also, I don't necessarily need the exact CLI program mentioned above. If some underlying tech exists and is easily scriptable on both platforms that would be great too.

    Read the article

  • Why does a redirect from a local IP address assume localhost?

    - by Jeremy
    I am developing a web application on my desktop and it is running on port 80. I am able to access the application from my laptop connected to the LAN by entering my desktop's LAN IP address 192.168.1.8. Now, my application sends a redirect after login, but my laptop assumes the final address is localhost/login. If I manually type in the IP address and URI for any page, it shows that I am logged in, so it works as expected. So, why does the redirect assume localhost? Both of my machines are linux-based. The laptop being Chrome OS. I am running nginx which proxies non-static file requests to jetty on port 8080.

    Read the article

  • PostgreSQL disaster recovery options

    - by Alex
    My customer has quite a large (the total "data" folder size is 200G) PostgreSQL database and we are working on a disaster recovery plan. We have identified three different types of disasters so far: hardware outage, too much load and unintentional data loss due to erroneously executed bad migration (like DELETE or ALTER TABLE DROP COLUMN). First two types seem to be easy to mitigate but we can't elaborate a good mitigation plan for the third type. I proposed to use ZFS and frequent (hourly) snapshots but "ZFS" means "OpenIndiana" these days and our Ops engineers do not have much expertise in it, so using OpenIndiana imposes another risk. Colleagues try to convince me that restoring from PostgreSQL PITR backup can be as fast as restoring from a ZFS snapshot but I highly doubt that replaying, say, 50G of archived WALs can be considered "fast". What other options are we missing? Is ZFS an only viable alternative? Can we get a fast Pg DB restore time in the Linux environment?

    Read the article

  • Webserver: Performance impact when storing session files on /dev/shm

    - by GetFree
    I have a website runing on a typical setup: Linux, Apache, PHP, MySQL. However, what's not typical about it, is that it's getting tons of traffic (400,000+ visits a day) and so, efficiency is becoming more and more important to me. I'm constantly looking for things I could optimize and, right now, my attention is focused on PHP's session files. There's a hell lot of session files constantly being read and created on the /tmp directory. So my question is: Is it a good idea to store the session files in /dev/shm (tmpfs) in order to speed things up a little bit??

    Read the article

  • How to get a maximum file size of VZFS parition?

    - by Nulldevice
    I have a VPS hosting with a VZFS file system. How can I determine maximum file size of VZFS partition? UPD: Free space (or total space) is not what i need. Sometimes file cannot occupy a hole partition volume - fat16 with 2Gb limit is a good example. I need to use a large database file (say, 64Gb) and so I need to know if a file system of VPS hosting will cope with it. It is easy to calculate for an ext3 filesystem using tune2fs, but VPS uses VSFS by Virtuozzo, and it is documented bad. Is it any generic way to calculate maximum file size for some filesystem in linux?

    Read the article

  • /data/tmp on database server?

    - by Mellon
    I am on a Linux ubuntu machine with MySQL installed. My teacher gave out an assignment which mentioned "copy cars.dat to /data/tmp on the MySQL database server" without any explanations, I do not know what is the "/data/tmp on database server" means exactly? Basically after that I need to execute SQL statement like LOAD DATA INFILE '/data/tmp/cars.dat' INTO TABLE cars So, what does copy cars.dat to /data/tmp on the database server means as there is no /data/tmp directory even? Personally, I checked /etc/mysql/my.cnf file, inside which there are definitions of : ... basedir = /usr datadir = /var/lib/mysql tmpdir = /tmp ... Does it mean to copy cars.dat to the tmpdir which is just /tmp under root directory??

    Read the article

  • DHCPOFFER delay VLAN

    - by john883
    I have configured two VLANs [ 15 and 16 ] and a trunkport on a Cisco Catalyst 2960. The trunkport is connected to eth2 on a Linux server The server is configured to support VLAN's and the interfaces eth2.15 eth2.16 is configured with ip addresses on two different subnet. dhcp3-server is running on the same server and hands out IP-addresses to the VLANs. When connecting a client to a port that is configured in ex. VLAN 15 and requesting a IP-address, i experience a long delay before recieving a DHCPOFFER, around 30 seconds or so, the client needs to send a DHCPDISCOVER about five times but will always recieve a DHCPOFFER. Any suggestion why this delay is happening?

    Read the article

  • Finding proof of server being compromised by Black Hole Toolkit exploit

    - by cosmicsafari
    I recently took over maintenance of a company server. (Just Host, C Panel, Linux server), theres a tonne of websites on it which i know nothing about. It had came to my attention that a client had attempted to access one of the websites hosted on this server and was met with a warning from windows defender. It had blocked access because it said the website had been compromised by the Black Hole Toolkit or something to that effect. Anyway I went in and updated various plugins and deleted some old suspect websites. I have since ran the website in question through a few online malware scanners and its comes up clean everytime. However im not convinced. Do any of you guys know extensive ways i can check that the server isn't still compromised. I have no way to install any malware scanners or anti virus programs on the server as it is horribly locked down by Just Host.

    Read the article

  • How useful is mounting /tmp noexec?

    - by Novelocrat
    Many people (including the Securing Debian Manual) recommend mounting /tmp with the noexec,nodev,nosuid set of options. This is generally presented as one element of a 'defense-in-depth' strategy, by preventing the escalation of an attack that lets someone write a file, or an attack by a user with a legitimate account but no other writable space. Over time, however, I've encountered arguments (most prominently by Debian/Ubuntu Developer Colin Watson) that noexec is a useless measure, for a couple potential reasons: The user can run /lib/ld-linux.so <binary> in an attempt to get the same effect. The user can still run system-provided interpreters on scripts that can't be run directly Given these arguments, the potential need for more configuration (e.g. debconf likes an executable temporary directory), and the potential loss of convenience, is this a worthwhile security measure? What other holes do you know of that enable circumvention?

    Read the article

  • monitor just what's going on -on a firewall

    - by bbutle01
    I have this little snapgear firewall. It's a little purpose built box running a custom linux, SH4 processor @240 Mhz, 64MB of ram. Basically how close we are to capacity is a mystery to me. I know I can run top and see the status of all the processes, but how can I see just how much of the processor is going to passing data... and how to estimate when I'm going to need to upgrade, and by tweaking iptables rules, how does that help/hurt the processor. suggestions?

    Read the article

  • Networking conflict - What is the most common default computer name?

    - by John
    I recently had to change the name of my computer to log onto a public wi-fi spot, because a computer with my name was already logged on. (I asked a guy there what to do after it said there was already a computer named "Dell(omitted)" logged on.) I've never been at a wifi spot that you had to log into before. I didn't even notice what the computer's name was before. My question is what are the most common default computer names. I'm curious. How often does this sort of problem happen? (I read via Google that Redhat linux default to "localhost" and Toshiba laptops default to "Toshiba.") Somebody probably keeps statistics. (I was referred here from serverfault.com) Thanks

    Read the article

  • In CentOS 4.3 Webmin 1.3000 bandwidth monitoring is eating disk space. How to delete those files?

    - by Silkograph
    I maintain Linux server being used for Mail, Squid and DNS service. Recently I observed that something was eating server disk space. But at last, today I caught the culprit which was consuming the disk by storing large number of files. On this server, Webmin 1.300 is installed. We use Squid proxy and Sarg to monitor Internet access. I always manually clear Sarg generated files under /var/www/html/squid for last few years. But I never realized that Webmin is also storing some kind of bandwidth log files in its' directory structure. I have noticed that under /etc/webmin/bandwidth/hours it has stored more thousands of files since year 2007 totaling about 17 GB. We have used 40 GB HDD for this server machine. My question is how can I delete those (/etc/webmin/bandwidth/hours) files safely?

    Read the article

  • Is it my motherboard?

    - by jarodrussell
    Early this week my home server (a Linux machine) threw a kernel panic. Yesterday is happened a couple of times. Then all of a sudden, when I plugged a USB stick in to run a memory test, the monitor stopped coming on. Now whenever I turn it on, the system gets power...I heard the drives spin, I see the processor fan spin, and the hard drive light comes on...but nothing happens. I put a video card in the AGP slot, but still nothing. The light on the power button that usually comes on stopped coming on. I took the memory out to see if it would beep, but nothing beeped. It's like it's getting power, but it's not coming on. Does this sound like a motherboard problem to anyone else? I think it is, but a second opinion would be greatly appreciated. Thanks.

    Read the article

  • setting up bridged adapter for VPN server

    - by B. VB.
    I have an Ubuntu linux Linode server that I am trying to install OpenVPN on. I'm following the tutorials (which, it turns out, are quite incomplete). auto br0 iface br0 inet static address 192.168.0.10 network 192.168.0.0 netmask 255.255.255.0 broadcast 192.168.0.255 gateway 192.168.0.1 bridge_ports eth0 bridge_fd 9 bridge_hello 2 bridge_maxage 12 bridge_stp off When I add this chunk in my /etc/network/interfaces, and I restart networking, my eth0 interface does not have an IP and I cannot get on the network (I need to use a buggy, slow, and annoying AJAX term to do damage repair). Why does adding this screw everything up? Any tips on how to set up this bridged adapter?? Thanks in advance!

    Read the article

  • What's more cost effective, Hosting your web startup on Foss or Windows?

    - by user37899
    Hi, Not coming from the windows world, I'm confused about licensing. I think my knowledge may be out of date. Before we gave up with windows web servers (IIS 2), we used to have to pay Client Access licence's. This worked out quite expensive. Is it cheaper to host 1000's of users on Windows than use Free open source software tools? http://serverfault.com/questions/124329/network-load-balancing-efficience-and-limits. This post suggests I can pay $15 a month, for unlimited users. I certainly hope that this is unbiased view, I am a professional and use the right technology for the right job. I hope i am not feeling the wrath of a windows (or linux for that matter) fanboy. Perhaps a Microsoft certified Licensing person can clear this up. Should i be recommending to startups windows servers and products over lamp? Cheers!

    Read the article

< Previous Page | 287 288 289 290 291 292 293 294 295 296 297 298  | Next Page >