Search Results

Search found 10023 results on 401 pages for 'manage processes'.

Page 183/401 | < Previous Page | 179 180 181 182 183 184 185 186 187 188 189 190  | Next Page >

  • NTFS disk mounted as fuseblk in ubuntu 12.10 is very slow and a lot of errors when rsync. Is that not a rare thing?

    - by Pablo Marin-Garcia
    I am having problems with a NTFS disk mounted as a fuseblk in my ubuntu 12.10 through external usb3. When I did a 1.1TB backup with rsync the speed was 1-2MB/s (wiht a ext4 disk speed was 70 MB/s before and after trying the NTFS disk). Also after one hour errors started to appear: rsync: write failed on "xxx": No such file or directory recv_files: "yyy" is a directory #but this file is a FILE not a dir ??!! .... As this is the first time I have mounted the NTFS in linux for heavy usage (the data would be used in windows afterwards), I would like to know if this kind of thinks are common o was only that something became unstable in my system and a simply restart would probably have solved it. This leads me to the these questions: Can I trust fuse for manage NTFS disks? Or is a problem of the NTFS tools in linux not yet totally stables for writing? Do people is still suffering from low performance with fuse-NTFS vs ext4 (in the past I have read about people complaining about this)?

    Read the article

  • Hardware asset management systems

    - by Dave
    I need to track a bunch of specialized testing tools, They are hardware devices used for testing other equipement. Each device has a serial number and is sent out for use in testing. Occasionally they break and have to be sent to the manufacture for repair. I'm looking for an open source application (preferably a webapp) to help manage them. Right now we're using Excel and it's not scaling as we get more tools. They aren't computers so all the standard IT asset management systems don't really fit the bill. I found h-harmony, but that project seems dead?

    Read the article

  • Monitoring and terminating a hanged process in Linux

    - by Yoav
    Hi, I'm writing a script that runs many simultaneous processes that run the "dig" command. Once in a while (relatively rare, but happens in every run since I run dig many times) the dig command hangs with 0% CPU. Therefore, my script never terminates. I've created a monitor process for each dig command I run, which terminates it after a while, but I was wondering if there isn't a simpler and more efficient way to run a process with a pre-determined "expiration date", i.e. if the process runs more then X seconds it gets a signal that terminates it. Thanks!

    Read the article

  • Screen startup apps

    - by stillinbeta
    I know that most people don't bother with things like screen anymore, but I happen to really like it, even in this GUI day and age. I still do most of my development from a BASH prompt, so it's extremely useful to me. What I'm wondering is what the easiest way is to start an instance of screen (stored in a shell script or .screenrc or somewhere else) so that it starts up with set commands already running in set windows. For example, I use a django test server, so I'd like one window to come up running "python manage.py runserver" and another blank, waiting for commands. The man page is wholly indecipherable. These old unix utilities can do quite nearly everything, so I'm sure this is possible, but I can't for the life of me figure out how. I

    Read the article

  • Slowdown after repeated hibernation in XP SP3

    - by John
    I have a Dell laptop w/dual core and 4 gigs of memory. Recently my machine has started to go off in never never land for 5 to 20 minutes at a time with a lot of hard drive churning. Eventually it comes back and everything operates properly. If I manage to have task manager up CPU usage shows no more then 3 to 5% during this process. I am on a corporate network and have very good virus protection in place. The problem seems to have increased in frequency since I let an update happen to IE8 and/or I started using hibernate more frequently. Any ideas or suggestions? TIA J

    Read the article

  • How to troubleshoot down terminal service/CITRIX sessions, when the process just won't terminate?

    - by Chad
    Running a CITRIX presentation server farm, version 4.5.6 on Windows 2003 sp2. In the CITRIX Access Management Console, I sometimes get a session that shows it's in a down state - but has none of the normal info associated with it (user name, applications, client name, idle time, etc...). It does say which servers it's on, so I check out that server's terminal services manager. I can see the down session but cannot reset it. I get: (Error 7024 - the requested operation cannot be completed because the terminal connection is currently busy processing a connect, disconnect, reset, or delete operation.) So I go to the task manager and look up the processes running under that session id. I see it's one of my published apps, but when I try to end the process - it simply does nothing and the process remains. Any way to get rid of these sessions without a server reboot?

    Read the article

  • Remote monitoring with screen capture

    - by Anonymous
    Is there anyway i can perodically take screenshots of a remote computer using nagios? I am experimenting with Nagios, and i am trying to explore different monitoring. So my question is apart from using nagios to monitor cpu usage, bandwidth utilization, uptime etc.. can i monitor my worker's productivity by checking what is he doing on his computer in a form of image output. Being able to monitor processes would not be of much help to me, as i only know if he or she is running firefox.exe for example he maybe using excessive use of firefox for facebook or other stuff but he claims he is troubleshooting and looking for solutions on forums. I saw a check_vnc script but i am unable to install the requsite vnc server anyone succesfully tried the vnc script care to share how to go about it? If not anyother way to try this?

    Read the article

  • Encoding Problem with Zend Navigation using Zend Translate Spanish in XMLTPX File Special Characters

    - by Routy
    Hello, I have been attempting to use Zend Translate to display translated menu items to the user. It works fine until I introduce special characters into the translation files. I instantiate the Zend_Translate object in my bootstrap and pass it in as a translator into Zend_Navigation: $translate = new Zend_Translate( array('adapter' => 'tmx', 'content' => APPLICATION_PATH .'/languages/translation.tmx', 'locale' => 'es' ) ); $navigation->setUseTranslator($translate); I have used several different adapters (array,tmx) in order to see if that made a difference. I ended up with a TMX file that is encoded using ISO-8859-1 (otherwise that throws an XML parse error when introducing the menu item "Administrar Applicación". <?xml version="1.0" encoding="ISO-8859-1"?> <!DOCTYPE tmx SYSTEM "tmx14.dtd"> <tmx version="1.4"> <header creationtoolversion="1.0.0" datatype="tbx" segtype="sentence" adminlang="en" srclang="en" o-tmf="unknown" creationtool="XYZTool" > </header> <body> <tu tuid='link_signout'> <tuv xml:lang="en"><seg>Sign Out</seg></tuv> <tuv xml:lang="es"><seg>Salir</seg></tuv> </tu> <tu tuid='link_signin'> <tuv xml:lang="en"><seg>Login</seg></tuv> <tuv xml:lang="es"><seg>Acceder</seg></tuv> </tu> <tu tuid='Manage Application'> <tuv xml:lang="en"><seg>Manage Application</seg></tuv> <tuv xml:lang="es"><seg>Administrar Applicación</seg></tuv> </tu> </body> </tmx> Once I display the menu in the layout: echo $this->navigation()->menu(); It will display all menu items just fine, EXCEPT the one using special characters. It will simply be blank. NOW - If I use PHP's UTF8-encode inside of the zend framework class 'Menu' which I DO NOT want to do: Line 215 in Zend_View_Helper_Navigation_Menu: if ($this->getUseTranslator() && $t = $this->getTranslator()) { if (is_string($label) && !empty($label)) { $label = utf8_encode($t->translate($label)); } if (is_string($title) && !empty($title)) { $title = utf8_encode($t->translate($title)); } } Then it works. The menu item display correctly and all is joyful. The thing is, I do not want to modify the library. Is there some kind of an encoding setting in either zend translate or zend navigation that I am not finding? Please Help! Zend Library Version: 1.11

    Read the article

  • Server Performance

    - by Burt
    We have a dedicated server that we use to stage websites (our test server). The performance of the server has become really bad and we regularly have to restart it. When performance is poor I have checked task manager for the processes and memory but everything looks OK. We use a content management system and it is always when using the admin section of this CMS that we notice the performance degrade which makes me think it may have something to do with DB calls the CMS is making. Does this sound viable? Any other sggestions of how I can go about testing this? Thanks in advance...

    Read the article

  • postgres memory allocation tuning 2

    - by pstanton
    i've got a Ubuntu Linux system with 12Gb memory most of which (at least 10Gb) can be allocated solely to postgres. the system also has a 6 disk 15k SCSI RAID 10 setup. The process i'm trying to optimise is twofold. firstly a single threaded, single connection will do many inserts into 2-4 tables linked by foreign key. secondly many different complex queries are run against the resulting data, using group by extensively. this part especially needs to be optimised. i have four of these processes running at once in order to make use of the quad core CPU, therefore there will generally be no more than 5 concurrent connections (1 spare for admin tasks). what configuration changes to the default Postgres config would you recommend? I'm looking for the optimum values for things like work_mem, shared_buffers etc. relevant doco thanks!

    Read the article

  • mdadm raid5 recover double disk failure - with a twist (drive order)

    - by Peter Bos
    Let me acknowledge first off that I have made mistakes, and that I have a backup for most but not all of the data on this RAID. I still have hope of recovering the rest of the data. I don't have the kind of money to take the drives to a recovery expert company. Mistake #0, not having a 100% backup. I know. I have a mdadm RAID5 system of 4x3TB. Drives /dev/sd[b-e], all with one partition /dev/sd[b-e]1. I'm aware that RAID5 on very large drives is risky, yet I did it anyway. Recent events The RAID become degraded after a two drive failure. One drive [/dev/sdc] is really gone, the other [/dev/sde] came back up after a power cycle, but was not automatically re-added to the RAID. So I was left with a 4 device RAID with only 2 active drives [/dev/sdb and /dev/sdd]. Mistake #1, not using dd copies of the drives for restoring the RAID. I did not have the drives or the time. Mistake #2, not making a backup of the superblock and mdadm -E of the remaining drives. Recovery attempt I reassembled the RAID in degraded mode with mdadm --assemble --force /dev/md0, using /dev/sd[bde]1. I could then access my data. I replaced /dev/sdc with a spare; empty; identical drive. I removed the old /dev/sdc1 from the RAID mdadm --fail /dev/md0 /dev/sdc1 Mistake #3, not doing this before replacing the drive I then partitioned the new /dev/sdc and added it to the RAID. mdadm --add /dev/md0 /dev/sdc1 It then began to restore the RAID. ETA 300 mins. I followed the process via /proc/mdstat to 2% and then went to do other stuff. Checking the result Several hours (but less then 300 mins) later, I checked the process. It had stopped due to a read error on /dev/sde1. Here is where the trouble really starts I then removed /dev/sde1 from the RAID and re-added it. I can't remember why I did this; it was late. mdadm --manage /dev/md0 --remove /dev/sde1 mdadm --manage /dev/md0 --add /dev/sde1 However, /dev/sde1 was now marked as spare. So I decided to recreate the whole array using --assume-clean using what I thought was the right order, and with /dev/sdc1 missing. mdadm --create /dev/md0 --assume-clean -l5 -n4 /dev/sdb1 missing /dev/sdd1 /dev/sde1 That worked, but the filesystem was not recognized while trying to mount. (It should have been EXT4). Device order I then checked a recent backup I had of /proc/mdstat, and I found the drive order. md0 : active raid5 sdb1[0] sde1[4] sdd1[2] sdc1[1] 8790402048 blocks super 1.2 level 5, 512k chunk, algorithm 2 [4/4] [UUUU] I then remembered this RAID had suffered a drive loss about a year ago, and recovered from it by replacing the faulty drive with a spare one. That may have scrambled the device order a bit...so there was no drive [3] but only [0],[1],[2], and [4]. I tried to find the drive order with the Permute_array script: https://raid.wiki.kernel.org/index.php/Permute_array.pl but that did not find the right order. Questions I now have two main questions: I screwed up all the superblocks on the drives, but only gave: mdadm --create --assume-clean commands (so I should not have overwritten the data itself on /dev/sd[bde]1. Am I right that in theory the RAID can be restored [assuming for a moment that /dev/sde1 is ok] if I just find the right device order? Is it important that /dev/sde1 be given the device number [4] in the RAID? When I create it with mdadm --create /dev/md0 --assume-clean -l5 -n4 \ /dev/sdb1 missing /dev/sdd1 /dev/sde1 it is assigned the number [3]. I wonder if that is relevant to the calculation of the parity blocks. If it turns out to be important, how can I recreate the array with /dev/sdb1[0] missing[1] /dev/sdd1[2] /dev/sde1[4]? If I could get that to work I could start it in degraded mode and add the new drive /dev/sdc1 and let it resync again. It's OK if you would like to point out to me that this may not have been the best course of action, but you'll find that I realized this. It would be great if anyone has any suggestions.

    Read the article

  • How to enable Hardware-assisted decoding in Ubuntu 9.10, specifically for Totem and/or Gnome MPlayer

    - by techtechmo
    Hi all, just wanted to ask about Hardware-assisted 720p video playback on an Intel Atom D510 machine, running on Ubuntu 9.10 Karmic Koala. When I use Boxee with Hardware Assisted Decoding activated, I am able to play 720p videos -- whether on the local harddisk drive or on a remote server, through LAN. However when I play the same 720p videos through Totem or through Gnome MPlayer, the resulting playback suffers from stuttering and slideshow-like slowdowns. Would it be possible to make 720p video playback on Totem and Gnome MPlayer more smooth given that my machine's processor in the Intel Atom D510? If yes, how? Boxee seems to manage, so I assume it should also be possible to tweak Totem and/or Gnome MPlayer to be able to do the same. As an added note, I the machine's OS is Linux Mint 8. Installed RAM is 2GB.

    Read the article

  • Need help configurating my Tomcat server without any WAR files

    - by gablin
    I just reinstalled my entire server, and now I can't seem to get my JSP-based website to work on Tomcat anymore. I use the same server.xml file, which worked perfectly before the reinstallation, but no longer. Here's the content of the server.xml file which worked before: <!--APR library loader. Documentation at /docs/apr.html --> <Listener className="org.apache.catalina.core.AprLifecycleListener" SSLEngine="on" /> <!--Initialize Jasper prior to webapps are loaded. Documentation at /docs/jasper-howto.html --> <Listener className="org.apache.catalina.core.JasperListener" /> <!-- JMX Support for the Tomcat server. Documentation at /docs/non-existent.html --> <Listener className="org.apache.catalina.mbeans.ServerLifecycleListener" /> <Listener className="org.apache.catalina.mbeans.GlobalResourcesLifecycleListener" /> <!-- Global JNDI resources Documentation at /docs/jndi-resources-howto.html --> <GlobalNamingResources> <!-- Editable user database that can also be used by UserDatabaseRealm to authenticate users --> <Resource name="UserDatabase" auth="Container" type="org.apache.catalina.UserDatabase" description="User database that can be updated and saved" factory="org.apache.catalina.users.MemoryUserDatabaseFactory" pathname="conf/tomcat-users.xml" /> </GlobalNamingResources> <!-- A "Service" is a collection of one or more "Connectors" that share a single "Container" Note: A "Service" is not itself a "Container", so you may not define subcomponents such as "Valves" at this level. Documentation at /docs/config/service.html --> <Service name="Catalina"> <!--The connectors can use a shared executor, you can define one or more named thread pools--> <!-- <Executor name="tomcatThreadPool" namePrefix="catalina-exec-" maxThreads="150" minSpareThreads="4"/> --> <!-- A "Connector" represents an endpoint by which requests are received and responses are returned. Documentation at : Java HTTP Connector: /docs/config/http.html (blocking & non-blocking) Java AJP Connector: /docs/config/ajp.html APR (HTTP/AJP) Connector: /docs/apr.html Define a non-SSL HTTP/1.1 Connector on port 8080 --> <Connector port="8080" protocol="HTTP/1.1" connectionTimeout="20000" redirectPort="8443" /> <!-- A "Connector" using the shared thread pool--> <!-- <Connector executor="tomcatThreadPool" port="8080" protocol="HTTP/1.1" connectionTimeout="20000" redirectPort="8443" /> --> <!-- Define a SSL HTTP/1.1 Connector on port 8443 This connector uses the JSSE configuration, when using APR, the connector should be using the OpenSSL style configuration described in the APR documentation --> <!-- <Connector port="8443" protocol="HTTP/1.1" SSLEnabled="true" maxThreads="150" scheme="https" secure="true" clientAuth="false" sslProtocol="TLS" /> --> <!-- Define an AJP 1.3 Connector on port 8009 --> <Connector port="8009" protocol="AJP/1.3" redirectPort="8443" /> <!-- An Engine represents the entry point (within Catalina) that processes every request. The Engine implementation for Tomcat stand alone analyzes the HTTP headers included with the request, and passes them on to the appropriate Host (virtual host). Documentation at /docs/config/engine.html --> <!-- You should set jvmRoute to support load-balancing via AJP ie : <Engine name="Standalone" defaultHost="localhost" jvmRoute="jvm1"> --> <Engine name="Catalina" defaultHost="localhost"> <!--For clustering, please take a look at documentation at: /docs/cluster-howto.html (simple how to) /docs/config/cluster.html (reference documentation) --> <!-- <Cluster className="org.apache.catalina.ha.tcp.SimpleTcpCluster"/> --> <!-- The request dumper valve dumps useful debugging information about the request and response data received and sent by Tomcat. Documentation at: /docs/config/valve.html --> <!-- <Valve className="org.apache.catalina.valves.RequestDumperValve"/> --> <!-- This Realm uses the UserDatabase configured in the global JNDI resources under the key "UserDatabase". Any edits that are performed against this UserDatabase are immediately available for use by the Realm. --> <Realm className="org.apache.catalina.realm.UserDatabaseRealm" resourceName="UserDatabase"/> <!-- Define the default virtual host Note: XML Schema validation will not work with Xerces 2.2. --> <!-- <Host name="localhost" appBase="webapps" unpackWARs="true" autoDeploy="true" xmlValidation="false" xmlNamespaceAware="false"> --> <!-- SingleSignOn valve, share authentication between web applications Documentation at: /docs/config/valve.html --> <!-- <Valve className="org.apache.catalina.authenticator.SingleSignOn" /> --> <!-- Access log processes all example. Documentation at: /docs/config/valve.html --> <!-- <Valve className="org.apache.catalina.valves.AccessLogValve" directory="logs" prefix="localhost_access_log." suffix=".txt" pattern="common" resolveHosts="false"/> --> <!-- </Host> --> <Host name="www.rebootradio.nu"> <Alias>rebootradio.nu</Alias> <Context path="" docBase="D:/services/http/rebootradio.nu" debug="1" reloadable="true"/> </Host> </Engine> </Service> </Server> The JSP site doesn't use any WAR files or anything like that; there's just a default.jsp in the specified folder D:/services/http/rebootradio.nu which loads the site. As I said, this configuration worked before, but now with the latest verion of XAMPP and Tomcat it doesn't work anymore. All I get is a 404 message saying The requested resource () is not available.

    Read the article

  • Remotely managing Hyper-V VMs from Windows 8 Client

    - by Vazgen
    Currently, I have a core Hyper-V Server hosting VMs for a domain controller and several domain-joined VDI infrastructure servers. The VMs are connected in that domain environment, but the remote management of the physical Hyper-V Server is set up using the same WORKGROUP (as the Windows 8 client I'm managing from) This makes it cumbersome to manage the VMs hosted on the physical server from my remote management Windows 8 client because I can only connect to the physical Hyper-V server and not the individual VMs hosted inside. Can I make my set up more flexible by hosting a second domain controller in a VM hosted on my Windows 8 machine and switching my remote management set up to use the same domain through? Meaning ALL physical and virtual machines including the VDI infrastructure under the same domain? I'm new to this just looking for some suggestions.

    Read the article

  • How do I silence strace's message "[ Process PID=15733 runs in 64 bit mode. ]" ?

    - by Ross Rogers
    I'm using memoize.py, but strace keeps injecting the following into the program output each time a process is executed: [ Process PID=15733 runs in 64 bit mode. ] or [ Process PID=16503 runs in 32 bit mode. ] How can I silence strace such that it doesn't inject these statements into the log file? At the very least, I'd like these statements to only go into the output file that memoize.py is instructing strace to use. It's already telling strace to put its output into a specific file ithrough arguments -o /tmp/OUTFILE. Note that strace is being called with the -f parameter to follow child processes.

    Read the article

  • Online Backup strategy for family individuals

    - by marlon brando
    The majority of my family, including myself, uses Dropbox and Syncplicity free accounts for our syncing needs. However our backup strategy is pretty non-existent, we all have access (via webDav) to our qnap nas located at home, however we copy files accross when we can remember to do so, terrible I know. Is there tools like CrashPlan or Spideroak that allows each the accomodation of each family members syncing and back-up needs under one main account? I'm not sure how this would work, as each family member would need a seperate sub-account or would I manage each computer's syncing and backup lists from a single account? Any thoughts?

    Read the article

  • Tips on Managing Podcast Subscriptions

    - by Ben Griswold
    I listen to a silly number of technical podcasts. I listen to enough of them that it is literally impossible to keep up. I nearly gave up and started dropping feeds from my subscription list when I heard Craig Shoemaker talk about his Polymorphic Podcast fast feed. The idea is he provides the same content at a higher speed so you can listen to his complete show in 3/4th the time. I tried it out with his recent jQuery Secrets with Dave Ward interview and I was shocked with the feed quality. It was a super clear, understandable conversation which only took a fraction of the time commitment. I experimented a bit and played the normal recording at 2x speed on my iPhone and the quality was once again just fine. But now I'm saving half of the time. I'm curious as to how you might manage your podcast subscriptions. Can you offer any tips or advice on how to get the best bang for your buck when it comes to technical podcast listening?

    Read the article

  • Excel: VLookUp solves a potential Nested IF problem.

    - by Mike
    I manually enter numbers on one cell according to text values in the cell adjacent to it. Is there a way to use the IF function to help me manage this? The text is automatically generated with a report but I put the numbers in manually in Excel. Example of my weekly boredom below: number Text in Cell 3 Order A 3 Order A 1 Order C 2 Order B 3 Order A 1 Order C 2 Order B 2 Order B HELP! My eyes and soul hurt each time I need to do this. Thanks Mike

    Read the article

  • IIS restart causes many Connection_Abandoned_By_AppPool messages in log

    - by PP
    After about half a day the IIS 6.0 ASP (VBScript) web server appears to die. It accepts TCP connections (i.e. I can telnet localhost 80 successfully) however it does not return any response. When I restart the server (going through My Computer - Manage - IIS - All Tasks - Restart) I get a lot of entries in C:\WINDOWS\system32\LogFiles\HTTPERR\httperr1.log at the time of restart that say: ... 80 HTTP/1.1 GET / - 843545307 Connection_Abandoned_By_AppPool DefaultAppPool ... 80 HTTP/1.0 GET / - 843545307 Connection_Abandoned_By_AppPool DefaultAppPool ... 80 HTTP/1.0 GET / - 843545307 Connection_Abandoned_By_AppPool DefaultAppPool where the ... refers to the prefix of the line, e.g. 2010-03-17 10:05:06 127.0.0.1 4927 127.0.0.1 80 There are 583 such lines (Connection_Abandoned_By_AppPool) in my log at time of restart. What could be the cause of this problem?

    Read the article

  • SQL Server: Is it possible to prevent SQL Agent from failing a step on error?

    - by franklinkj
    I have a stored procedure that runs custom backups for around 60 SQL servers (mixes 2000 through 2008R2). Occasionally, due to issues outside of my control (backup device inaccessible, network error, etc.) an individual backup on one or two databases will fail. This causes this entire step to fail, which means any subsequent backup commands are not executed and half of the databases on a given server may not be backed up. On the 2005+ boxes I am using TRY/CATCH blocks to manage these problems and continue backing up the remaining databases. On a 2000 server however, for example, I have no way to prevent this error from failing the entire step: Msg 3201, Level 16, State 1, Line 1 Cannot open backup device 'db-diff(\PATH\DB-DIFF-03-16-2010.DIF)'. Operating system error 5(Access is denied.). Msg 3013, Level 16, State 1, Line 1 BACKUP DATABASE is terminating abnormally. I am simply asking if anything like this is possible in SQL 2000 or if I need to go in a completely different direction.

    Read the article

  • Best CMS to handle short snippets of text?

    - by Federico Poloni
    I have to install a CMS to manage a set of mathematical problems, i.e., our main content will be short (~3 lines) snippets of text. We need the ability to add comments and categories/tags, possibly with a powerful search function combining different constraints on the categories. A crucial ability is the possibility to combine the results of a search in the same page to produce a (printable) problem sheet: not many CMS's seem to be able to do so, and it is difficult for me to test every one for this specific function. Do you guys know of a CMS that is capable to return formatted search result in this fashion? Thanks in advance!

    Read the article

  • Trying to understand why VLANs need to be created on intermediate switches

    - by Jon Reeves
    I'm currently studying for the Cisco switching exam and having trouble understanding exactly how 802.1q tagging works. Given three daisy chained switches (A,B, and C) with trunk ports between them and VLAN 101 defined on both end switches (A and C), I'm not sure why the VLAN also needs to be defined on the middle one (B)? Note that I am not disputing that it does need to be configured, I'm just trying to understand why exactly. As I understand it, traffic from VLAN 101 on switch A will be tagged as it goes through the trunk to switch B. According to the documentation I have read, trunks will pass all VLANs by default, and the .1q tag is only removed when the frame leaves through an access port on the relevant VLAN. From this I would expect switch B to simply forward the tagged frame unchanged through the trunk to switch C. Can anyone shed some light on how switch B processes this frame and why it does not get forwarded through the other trunk ?

    Read the article

  • Is it possible to format ps RSS (memory) output to be more human friendly?

    - by metasoarous
    Executing ps ux returns a nice list of process information, easy to grep through or watch. However, there doesn't seem to be much flexibility in the memory usage output; the RSS (resident set size) is printed in kB, which for large processes is hard to read (especially at a glance), and %MEM gives 100 × RSS / system_memory. The du utility has a lovely -h flag which prints space in a more user friendly fashion. I have not been able to find anything equivalent for ps. Is there a special formatting trick that can accomplish this?

    Read the article

  • Are mobo raid controllers based on Intel ICH10R southbridge still considered software Raid?

    - by Breadtruck
    So do you still need specific software/drivers (intel matrix?) installed to manage the raid controller, setup the array. If the raid chipset is on-board the motherboard and it uses the CPU, and say I am using a Core 2 Quad Q9550, would a hardware based card still out perform the motherboard chipset? This is for a home workstation and I could spend $300 on a areca ARC-1210 PCI-Express x8 SATA II but I want to be able to justify the money for the raid card. My motherboard is a GIGABYTE GA-EP45-UD3P UPDATE: I was going to RAID5 using 4 500 GB drives, and I was going to buy a controller card but this article got me thinking hmmm....Toms-Southbridge Battle

    Read the article

  • Skype companywide global contacts list

    - by Martin
    We are a medium sized company based across several sites and with a number of home workers. We have more or less settled on Skype as our defacto method of communication. At the moment the only pain is ensuring that everybody has all the other employees added to their contact list. Can be a real pain when a new employee starts and they have to send details to everyone else and vice versa. Is there a solution that allows us to manage a central contacts list that we can push out to new/existing users?

    Read the article

< Previous Page | 179 180 181 182 183 184 185 186 187 188 189 190  | Next Page >