Search Results

Search found 17856 results on 715 pages for 'setup py'.

Page 404/715 | < Previous Page | 400 401 402 403 404 405 406 407 408 409 410 411  | Next Page >

  • BIND9 DNS Server

    - by Stevecomptech
    I am trying to set up my own DNS server solaris 10, i used the setup guide from http://www.logiqwest.com/dataCenter/Demos/RunBooks/DNS/DNSsetup.html I tried to join windows xp clients to the domain i configured using the above guide but i was unable to join giving me these errors, An error occurred when DNS was queried for the service location (SRV) resource record used to locate adomain controller for domain MYDOMAIN.COM. The error was: "No records found for given DNS query." (error code 0x0000251D DNS_INFO_NO_RECORDS) The query was for the SRV record for _ldap._tcp.dc._msdcs.MYDOMAIN.COM Where could i be going wrong? Your assistance will highly be appreciated

    Read the article

  • VirtualBox - Mac OSX host Win7 guest - no Internet access for guest VM

    - by nodelayheehoo
    I have a Mac running OSX 10.9.2, and I just downloaded and installed a Win7 IE9 VM in VirtualBox. My Mac uses Wi-Fi for internet access, and it's behind a proxy (it's a work machine). VirtualBox loads the VM fine, and at some point the VM can see the DNS servers of the host. But I've never been able to make the VM have internet access. I've tried all kinds of combinations of Network settings on the VM via the VirtualBox Settings, in conjunction with Internet Sharing in OSX's System Preferences, but no luck. Has anyone done a similar setup and made the VM successfully connect to the Internet? Thanks in advance for any inputs. [ Update: I was able to get internet access for the VM when the host was using my home network. When I ran the VPN software to connect to the work network, the internet access went away again.] (Initially posted this on stackoverflow.com, but it was put on hold as off-topic by several users, and was advised to ask here instead)

    Read the article

  • WMIC returns error when querying product

    - by Stu
    I'm trying to automate the installation of an MSI on my server, however before the installation can go ahead I need to uninstall the previous version from the server. Searching on the internet I've found that WMIC is the tool required but there seems to be a problem with the setup of WMI on the server. Running the following command gives errors: command promptwmic then inside the tool /trace:on product get name This returns a long string of successes and one failure: FAIL: IEnumWbemClassObject->Next(WBEM_INFINITE, 1, -, -) Line: 396 File: d:\nt\admin\wmi\wbem\tools\wmic\execengine.cpp Node - ENTECHORELDEV ERROR: Code = 0x80041010 Description = The specified class is not valid. Facility = WMI I'm trying to run this on a standard install of Windows Server 2003 R2 with administrator privelages. Thanks Stu

    Read the article

  • A question about the cobbler-ubuntu-import bash script

    - by user183394
    I have been testing the latest cobbler for PXE booting Ubuntu 12.04.1-server-x86_64 and 12.10-server-x86_64 using a Scentific Linux 6.3 host to run the cobbler server. With the former, I got everything going. But with the later, I haven't been successful. As an attempt to figure things out, I downloaded Ubuntu's cobbler 2.2.2 source package. Examining the content, I soon noticed that Ubuntu's cobbler 2.2.2 came with a cobbler-ubuntu-import bash script. I reviewed the code and spotted something interesting: line 9 of the script states: 9 AUTO_KOPTS='log_host=@@server@@ log_port=514 priority=critical locale=en_US netcfg/choose_interface=auto' But after extensive googling, reading both Debian and Ubuntu's documentation about the debian-installer, I don't see these two kopts log_host and log_port documented anywhere. Putting it in the profile of my current test setup, even my cobbler server host does run rsyslogd, I don't see anything logged either. No, I don't have iptables and selinux on on the cobbler server host. Can anyone point to me where I can read more about these two options? Having the ability to log an installation to a remote central logging host would be really cool.

    Read the article

  • php mail not arrives at gmail, not at local server

    - by thomas
    The php mail function I am using does not work completely. It will sent mails to gmail easy enough. However, emails routed directly to my internally hosted exchange server are not getting through. The servers/domains are setup is as follows. URLs are registered with Network solutions (www.independentsservice.com & www.isco.net) NS directs all traffic to our ISP (Socket.net). Socket directs as follows: Mail to our local server FTP to our local server HTTP to our website hosted on Chihost.com Traffic to our local server goes through a Watchguard firewall which routes mail traffic to our locally hosted Exchange server. Is there some reason why exchange won't accept these emails? Thanks!

    Read the article

  • E-Business Suite - Cloning Basics & AMP Cloning - EMEA/APAC

    - by Annemarie Provisero
    ADVISOR WEBCAST: E-Business Suite - Cloning Basics & AMP Cloning - EMEA/APAC PRODUCT FAMILY: EBS – ATG - Utilities July 19, 2011 at 10:00 am CET, 01:30 pm India, 05:00 pm Japan, 06:00 pm Australia This 1.5-hour session is recommended for technical and functional Users who are interested to get an generic overview about the Cloning functionality available in the E-Business Suite Release. We are going to talk about the generic Cloning options and will then go into depth about the cloning scenario when using AMP (Applications Management Pack) within the Enterprise Manager. TOPICS WILL INCLUDE: Cloning Overview Rapidclone steps in Details Rapidclone limitations EM Grid Setup with AMP for Cloning Advantages of Cloning with AMP Cloning Procedures available with AMP Monitoring Clone Operation Few things to remember before Cloning A short, live demonstration (only if applicable) and question and answer period will be included. Oracle Advisor Webcasts are dedicated to building your awareness around our products and services. This session does not replace offerings from Oracle Global Support Services. Click here to register for this session ------------------------------------------------------------------------------------------------------------- The above webcast is a service of the E-Business Suite Communities in My Oracle Support. For more information on other webcasts, please reference the Oracle Advisor Webcast Schedule.Click here to visit the E-Business Communities in My Oracle Support Note that all links require access to My Oracle Support.

    Read the article

  • Load balance between two virtual directories on a single machine running IIS6

    - by Emil Lerch
    I'm trying to simulate a production environment using a single VM. The production environment will have 3 servers behind a load balancer with sticky sessions. Production will be using IIS6. For development, I have no need to distribute load, but I want to make sure the team does not write code that relies on a single-server module. I have been hoping to set up two virtual directories and find some isapi filter to perform some rudimentary software load balancing, but I've been unable to locate anything suitable. isapi_rewrite comes close but there's no ability for a random rewrite. I don't even need sticky sessions since the application does not (in theory) rely on them. Is there something easy/free to do this, or am I stuck with multiple VMs and a NLB setup (which seems like total overkill)?

    Read the article

  • Mini-DisplayPort to HDMI Adapter Sound Stopped Working

    - by jimdrang
    I have cancelled cable and want to watch the NCAA tournament games on my TV tonight through my 2011 Macbook Pro where I can stream the game in a browser. I have a cheap Mini-DisplayPort to HDMI converter that I have connected to my TV in the past and had no issues with audio or video, the problem is the audio has stopped working since the last time I used it a few months ago and now just keeps playing through the laptop speakers, but the video works fine. Everything with my setup is the same and when I try to force the audio output to the TV in the Audio system settings, my TV is not listed as an output option at all. I have tried various combinations of power cycling, replugging-in both devices and making sure the TV options are set properly to receive audio through HDMI but no luck. Anyone know what the issue could be?

    Read the article

  • Prevent Win7 boot loader from taking over the WinXP boot loader

    - by Chris
    My setup: 1 physical hard drive (500gb divided equally into 2 partitions) Windows XP Partition (Current OS) Empty Partition where I will be installing Windows 7 My question is how do I prevent the Windows7 boot loader from taking over my WindowsXP boot loader when installing the new OS ? The reason I am asking is because I already have a ghosted backup of my WinXP partition and if I ever need to restore my xp partition using that backup, would it not overwrite the Windows7 boot loader that was placed in the XP partition with the one from the backup, thus making windows 7 unable to boot. Also what would happen if I decided to delete the Windows XP partition altogether somewhere down the road and along with it the Win7 boot loader that was placed there, wouldn't that cause the system not to boot at all.. To avoid these issues, I simply want to make sure that BOTH the Win7 and WinXP boot loaders are available on their respective partitions and they do not interfere with each other in any way. Is this possible? Thx, Chris

    Read the article

  • How do internal smtp servers send mail to internal mail servers on the network?

    - by dmr83457
    We have upgraded our internal corporate email server and the IP address has changed internal to our network. A second email server is used for sending bulk jobs for a mailing list service that we offer. Since the switch of the internal corporate server IP we have been seeing problems when the bulk email server is trying to send email to our own domain. The log shows that it is still trying to hit the old corporate server instead of the new one. I have looked through all settings for the bulk email software and see nothing set there to send to internal mail servers, and I do not recall doing anything special to get this working when setup a couple years ago. Is there a setting in Win2003 or on the network that enables the mapping of MX record external IPs to internal IPs so mail gets routed correctly?

    Read the article

  • Google Analytics async=true seems wrong in the Google documentation?

    - by leeand00
    In the Google Analytics async example, they state that in order to include more than one tracker, you need to setup your pages for asyncrous tracking, and they do so using the following code: <script type="text/javascript"> _gaq.push( ['_setAccount', 'UA-XXXXX-1'], ['_trackPageview'], ['b._setAccount', 'UA-XXXXX-2'], ['b._trackPageview'] ); (function() { var ga = document.createElement('script'); ga.type = 'text/javascript'; ga.async = true; ga.src = ('https:' == document.location.protocol ? 'https://ssl' : 'http://www') + '.google-analytics.com/ga.js'; var s = document.getElementsByTagName('script')[0]; s.parentNode.insertBefore(ga, s); })(); </script> The second tracker is not receiving any results. After looking at my tracking codes to ensure they are correct, I noticed that the ga.async = true statement is specified differently most of the time and is never set to a value of true, it's often set to async but never true. Could this be stopping my Analytics data from posting to the second tracker? or might it be something else? Also what calls should I look for in the Net tab in firebug to ensure that GA is being called when the page loads?

    Read the article

  • Autocomplete in Silverlight with Visual Studio 2010

    - by Sayre Collado
    Last week I keep searching on how to use the autocomplete in silverligth with visual studio 2010 but most of the examples that I find they are using a textbox or combobox for the autocomplete. I tried to study those examples and apply to the single autocomplete from tools on my silverlight project. And now this is the result. I will use a database again from my previous post (Silverlight Simple DataBinding in DataGrid) to show how the autocomplete works with database. This is the output: First, this is the setup for my autocomplete: //The tags for autocompletebox on XAML Second, my simple snippets: //Event for the autocomplete to send a text string to my function private void autoCompleteBox1_KeyUp(object sender, KeyEventArgs e) { autoCompleteBox1.Populating += (s, args) => { args.Cancel = true; var c = new Service1Client(); c.GetListByNameCompleted +=new EventHandler(c_GetListByNameCompleted); c.GetListByNameAsync(autoCompleteBox1.Text); }; } //Getting result from database void c_GetListByNameCompleted(object sender, GetListByNameCompletedEventArgs e) { autoCompleteBox1.ItemsSource = e.Result; autoCompleteBox1.PopulateComplete(); } The snippets above will show on how to use the autocompleteBox using the data from database that bind in DataGrid. But what if we want to show the result on DataGrid while the autocomplete changing the items source? Ok just add one line to c_GetListByNameCompleted void c_GetListByNameCompleted(object sender, GetListByNameCompletedEventArgs e) { autoCompleteBox1.ItemsSource = e.Result; autoCompleteBox1.PopulateComplete(); dataGrid1.ItemsSource = e.Result; }

    Read the article

  • What are some of the benefits of a "Micro-ORM"?

    - by Wayne M
    I've been looking into the so-called "Micro ORMs" like Dapper and (to a lesser extent as it relies on .NET 4.0) Massive as these might be easier to implement at work than a full-blown ORM since our current system is highly reliant on stored procedures and would require significant refactoring to work with an ORM like NHibernate or EF. What is the benefit of using one of these over a full-featured ORM? It seems like just a thin layer around a database connection that still forces you to write raw SQL - perhaps I'm wrong but I was always told the reason for ORMs in the first place is so you didn't have to write SQL, it could be automatically generated; especially for multi-table joins and mapping relationships between tables which are a pain to do in pure SQL but trivial with an ORM. For instance, looking at an example of Dapper: var connection = new SqlConnection(); // setup here... var person = connection.Query<Person>("select * from people where PersonId = @personId", new { PersonId = 42 }); How is that any different than using a handrolled ADO.NET data layer, except that you don't have to write the command, set the parameters and I suppose map the entity back using a Builder. It looks like you could even use a stored procedure call as the SQL string. Are there other tangible benefits that I'm missing here where a Micro ORM makes sense to use? I'm not really seeing how it's saving anything over the "old" way of using ADO.NET except maybe a few lines of code - you still have to write to figure out what SQL you need to execute (which can get hairy) and you still have to map relationships between tables (the part that IMHO ORMs help the most with).

    Read the article

  • links for 2011-02-01

    - by Bob Rhubart
    OTN Virtual Developer Day for WebLogic Server and WebLogic Developer Broadcasts (WebLogic Server) Mike Lehmann with details on a whole bunch of upcoming online events for those with an interest in WebLogic. (tags: WebLogic oracle otn) IOUC Summit: Open Arms and Cheese Shoes (Oracle Technology Network Blog (aka TechBlog)) Event highlights from OTN head honcho Justin Kestelyn. (tags: oracle otn IOUC) Prognostications for the Future of BI (BI & Analytics Pulse) Jacqueline Coolidge looks into the Business Intelligence crystal ball. (tags: oracle otn businessintelligence) Edwin Biemond: Some handy code for your managed Beans ( ADF & JSF ) "Back in 2009, I already a made a blogpost about some handy code which you can use in your ADF Web Application. You can say this blogspot is part 2 and here I will show you the code, I use most in my own managed Beans." - Oracle ACE Edwin Biemond (tags: java SOA oracle oracleace) Leon Smiers: Process, content and collaboration "Taking a look at today’s business, most companies still have a lot [to do] as far as adapting to and leveraging Web 2.0 possibilities is concerned." - Leon Smiers (tags: e20 oracle enterprise2.0) Antony Reynolds: Using the SOA-BPM VIrtualBox Appliance Antony says: "Recently I have been setting up some machines for fellow engineers. My base setup consists of Oracle Enterprise Linux with Oracle Virtual Box." (tags: oracle otn soa virtualization virtualbox bpm) Oracle Weblogic Server Gets Smart with CERN | SiliconANGLE CERN, the home to European particle physics, chose Oracle Weblogic Server to handle technical applications and copious HR and administrative Java-based web applications used by CERN employees. Oracle got its start by scheduling the interventions of the Large Hadron Collider (LHC). (tags: Weblogic oracle CERN) Oracle Virtual Developers Day: WebLogic - February 10, 2011 Virtual Developer Day: WebLogic - February 10, 2011. Speakers: Frances Zhao - Principal Prod Mngr, Java Platform Group; Will Lyons - Dir, WebLogic Server Prod Mgmt; Steven Button - Principal Prod Mngr, WebLogic Server; Pyounguk Cho - Principal Prod Mngr, Java Platform Group. (tags: oracle otn weblogic java)

    Read the article

  • Term for unit testing that separates test logic from test result data

    - by mario
    So I'm not doing any unit testing. But I've had an idea to make it more appropriate for my field of use. Yet it's not clear if something like this exists, and if, how it would possibly be called. Ordinary unit tests combine the test logic and the expected outcome. In essence the testing framework only checks for booleans (did this match, did the expected result result). To generalize, the test code itself references the audited functions, and also explicites the result values like so: unit::assert( test_me() == 17 ) What I'm looking for is a separation of concerns. The test itself should only contain the tested logic. The outcome and result data should be handled by the unit testing or assertion framework. As example: unit::probe( test_me() ) Here the probe actually doubles as collector in the first run, and afterwards as verification method. The expected 17 is not mentioned in the test code, but stored or managed elsewhere. How is this scheme called? Or how would you call it? I hope I can find some actual implementations with the proper terminology. Obviously such a pattern is unfit for TDD. It's strictly for regression testing. Also obviously, it cannot be used for all cases. Only the simpler test subjects can be analyzed that way, for anything else the ordinary unit test setup and assertion steps are required. And yes, this could be manually accomplished by crafting a ResultWhateverObject, but that would still require hardwiring that to the test logic. Also keep in mind that I'm inquiring for use with scripting languages, and not about Java. I'm aware that the xUnit pattern originates there, and why it's hence as elaborate as it is. Btw, I've discovered one test execution framework which allows for shortening simple test notations to: test_me(); // 17 While thus the result data is no longer coded in (it's a comment), that's still not a complete separation and of course would work only for scalar results.

    Read the article

  • LCD? LED? Plasma? The How-To Geek Guide to HDTV Technology

    - by Eric Z Goodnight
    With image technology progressing faster than ever, High-Def has become the standard, giving TV buyers more options at cheaper prices. But what’s different in all these confusing TVs, and what should you know before buying one? If you’re considering buying a television this Holiday season for a loved one (or simply for yourself), it can be a big help to know what to look for. Take a look to find out what sets HD televisions apart, learn some of the confusing jargon associated with them, and see a comparison of four of the types of HDTVs commonly sold today. Latest Features How-To Geek ETC The How-To Geek Guide to Learning Photoshop, Part 8: Filters Get the Complete Android Guide eBook for Only 99 Cents [Update: Expired] Improve Digital Photography by Calibrating Your Monitor The How-To Geek Guide to Learning Photoshop, Part 7: Design and Typography How to Choose What to Back Up on Your Linux Home Server How To Harmonize Your Dual-Boot Setup for Windows and Ubuntu Hang in There Scrat! – Ice Age Wallpaper How Do You Know When You’ve Passed Geek and Headed to Nerd? On The Tip – A Lamborghini Theme for Chrome and Iron What if Wile E. Coyote and the Road Runner were Human? [Video] Peaceful Winter Cabin Wallpaper Store Tabs for Later Viewing in Opera with Tab Vault

    Read the article

  • Dock with dual external DVI monitors with Intel + Nvidia Optimus?

    - by Ryan
    I have a Dell Latitude E6420 laptop plugged into a docking station, and the dock has 2 monitors (connected with DVI). Also note that I've installed Ubuntu alongside (dual-boot) Windows 7. I can't get the dual monitors to work both on Ubuntu (either 11.10 or 12.04) and Windows 7. When I run lspci | grep VGA, I get: 00:02.0 VGA compatible controller: Intel Corporation 2nd Generation Core Processor Family Integrated Graphics Controller (rev 09) 01:00.0 VGA compatible controller: nVidia Corporation GF108 [Quadro NVS 4200M] (rev a1) If I then reboot and uncheck Optimus setting in the BIOS during reboot, I'm able to get the dual monitors to work in Ubuntu 12.04 (but I need to configure them every boot in Nvidia Settings). When I run lspci | grep VGA, I get: 01:00.0 VGA compatible controller: NVIDIA Corporation GF119 [Quadro NVS 4200M] (rev a1) But then if I reboot into Windows (leaving the Optimus unchecked), Windows can't detect external monitors, and the resolution is unacceptably low. I've seen on many forum posts that this particular graphics card setup causes lots of headaches. I haven't been able to resolve my problem yet. How can I use my external display on my laptop with intel and nvidia video cards? How to use external displays with Intel driver on a NVidia/Intel hybrid system nVidia Optimus , Unity 3D and Dual Monitors "Just use VGA instead of DVI" isn't an option because my dock has only 1 VGA port (and 2 DVI). Switching the BIOS setting on every reboot and then reconfiguring the display settings every time is tedious, time-consuming, and impractical. Do you know how to make this work smoothly? Thanks for your help! P.S. see also: http://superuser.com/questions/434358/dell-latitude-e6420-dual-boot-ubuntu-windows-7-optimus-graphics-problems

    Read the article

  • Installing Gitolite on NAS - FATAL: have errors but logging failed

    - by Jay
    I'm trying to install Gitolite on my Synology DiskStation, following these instructions: http://www.bluevariant.com/2012/05/comprehensive-guide-git-gitolite-synology-diskstation/ Under "Install Gitolite on the DiskStation and Run Setup" When I run the gitolite install command: DiskStation> /volume1/homes/git/gitolite/install -ln /bin I get the following error: FATAL: have errors but logging failed! 2012-05-31.00:10:22 no GL_LOGFILE env var 2012-05-31.00:10:22 die could not symlink /volume1/home/git/gitolite/src/gitolite to /bin<<newline>> at /volume1/home/git/gitolite/install line 71<<newline>> I am very new to all of this. Does anyone know what this means, and how I can fix this error? Thank you very much.

    Read the article

  • How to install apt-get on a busybox embedded system?

    - by Daniel YC Lin
    My embedded system is for sh4 CPU. The debian distribution may get on http://www.si-linux.co.jp/pub/debian-sh/lenny-sh4/ I get the apt*.deb and extract the data.tar.gz. After setup the /etc/apt/sources.list, I could do 'apt-get update'. But it missing dependency when I try to run 'apt-get install ntpdate'. Is there any method to let apt-get ignore some base packages? Because those package is build by my original embedded system.(eg. busybox).

    Read the article

  • bash code in rc.local not excuting after bootup

    - by mrTomahawk
    Does anyone know why a system would not execute the script code within rc.local on bootup? I have a post configuration bash script that I want to run after the initial install of VMware ESX (Red Hat), and for some reason it doesn't seem to execute. I have the setup to log its start of execution and even its progress so that I can see how far it gets in case it fails at some point, but even when I look at that log, I am finding that didn't even started the execution of the script code. I already checked to see that script has execution permissions (755), what else should I be looking at? Here is the first few lines of my code: #!/bin/sh echo >> /tmp/configLog "" echo >> /tmp/configLog "Entering maintenance mode"

    Read the article

  • mutt, smime, decrypt with one of two different keys

    - by munin
    This is an odd one. We want to have an encrypted e-mail list. There are a few ways to do this, but in the interim what we've done is created a public/private keypair via openssl for our e-mail list ([email protected]) and then distributed the public/private keypair amongst the list participants (ugh). When someone posts to the list, they encrypt using the lists public key, and everyone has the private key (ugh) so it 'works'. MUAs like Outlook and Thunderbird work with this setup. Mutt has a problem though - it seems to only decrypt a SMIME message with a private key that is specified by your e-mail. So when someone sends an e-mail to the list e-mail, my MUA won't decrypt it. How can I tell Mutt about this second private key?

    Read the article

  • A little tidbit on Team Build 2010 and error MSB3147

    - by Enrique Lima
    The problem? Performing a build on a ClickOnce solution would not be successful due to the setup.bin not being located. Ok, now what? Researched from corner to corner, install, re-install, update.  Found some interesting posts to fix the issue, but most of them were focused on Team Foundation Server/Team Build 2008, and some other on 2005.  The other interesting tidbit was the frequent indication to modify the registry to help Team Build find the bootstrapper. Background info:  This was a migration I posted about a few days ago, a 32 bit TFS implementation to a full 64 bit TFS implementation.  Now, the project has binaries and dependencies on X86 (This piece of information became essential to moving from a failed build to a successful build). So, what’s the fix? The trick in this case was to go back into the Build Type and check the properties/configuration.  Upon further investigation, I found the following:  Once you Edit the Build Definition, then select Process, expand 3. Advanced and look for MSBuild Platform, switch from Auto to X86.  Ran the Build, and success!

    Read the article

  • Ubuntu not shutting down ( going to black screen ) 12.04

    - by Orrin Fox
    I am currently using a USB persistent install of ubuntu. its a simple 4GB drive with a 2.8GB partition ( casper-rw storage partition ). I setup an administrator account and set it to login automatically. I also removed ubiquity to simply use this as a go anywhere install. Heres my issue. Im logged in as my account, and I click the top right gear and select "shut down". Text pops up showing its quitting processes.. etc. and then goes to the plymouth animation. But... The screen goes black, and then it goes to the login screen. Now when im at the login screen i go into terminal ( alt+F2 ) and dont you know, im logged in as Ubuntu. so then I try the following: ubuntu@ubuntu:~$ sudo shutdown now It goes to the plymouth screen again as if its shutting down, AND the screen goes black once again but the computer has not turned off, as in the usb is still flashing the light, the fans are still on, the only thing off is the screen. Is this a bug? If not maybe i did something wrong? Perhaps its that I made an account but... if there is a work around for this please let me know. Thanks again, Fox

    Read the article

  • SNMPD - passing values in a trap

    - by ethrbunny
    Im trying to create an SNMP setup to monitor 500-1K machines. Each will be joined to the SNMPTRAPD server via a VPN. When the client system sends its messages, it's sending its internal address instead of the VPN address even though the 'trapsync' command is using the VPN. This will be a problem as many of these system will have DHCP addresses based on location so there will be overlap. What I'd like to figure out is how to pass extra values with the trap so I can identify which system a given trap comes from.

    Read the article

  • Unable to access internal network through PfSense WAN port

    - by Sean
    Our branch office is unable to connect to our internal network for some reason. However we can connect to the branch office domain controller from behind PfSense. The following is our setup: |Branch DC - 192.168.0.101 | |Branch Firewall - 192.168.0.2 | |(Internet) | |Local Firewall - 192.168.3.1 | |PFSense WAN port - 192.168.3.100 |PFSense LAN port - 192.168.1.1 | | DC1 - 192.168.1.2|DC2 - 192.168.1.4 Branch DC can ping and connect to PFSense WebGUI on the WAN port successfully (we set this up using the PfSense documentation). DC1 and DC2 can connect outbound to the Branch DC. There seems to be a rule that prevents internal access on the WAN port. However our rules are set to allow all traffic on the LAN and WAN ports. It would be ideal to just disable the firewall altogether since we already have a firewall but when we do this, PfSense doesn't allow any internal traffic at all. I look forward to any assistance and thank you ahead of time.

    Read the article

< Previous Page | 400 401 402 403 404 405 406 407 408 409 410 411  | Next Page >