SAP BusinessObjects (BO) BI 4 Platform has come with many improvements that have made the system more complex. Rep-Up-GraphWe, as BI 4 administrators, are required to know how this application works in order to apply the proper optimization improvements to the platform, when it is required. In this blog article, we will cover the most important concepts that need to be considered as a first step to increase the performance of the BI platform. It is very important for a good designer to understand the conceptual tiers in the BI Platform: Application Tier, Intelligence tier and Processing tier. We will refer to each one of them and we will nail down some optimization tips.
Running SAP Explorer on top of BW This blog describes the only successfully tested way to run SAP Explorer on top of SAP Netweaver Business Warehouse (SAP BW) info-providers. This is not a straight forward configuration and it requires knowledge of Information Design Tool, Data Federator Administration Tool and SAP Logon application for SAP BW.
The supported connection by SAP Explorer to explore data on top of SAP BW info-providers is through a Relational Universe using a JCo (Java Connector). But this connection does not work with just configuring the right parameters for the JCo connector in IDT, it requires some additional adjustments to be done on the SAP BusinessObjects - SAP BW environment.
Here is a quick brief on some topics to be considered (how-to not explained in this blog post):
- It is a recommendation from SAP to split (size) the Adaptive Processing Server in up to 6 different APSs and isolate this way some of the services (SAP KBA 1694041)
- Make sure Data Federation Service in included as a service in your current APS
- Make sure the user you are using for creating the relational connection is propertly granted in SAP BW (SAP KBA 1604227)
- Make sure with your IT team that there are no firewall activities or blocked ports that can make the connection fail
Step 1: Creating a Relational Connection in IDT:
Open IDT, create a new Relational Connection - SAP - SAP NetWeaver BW - SAP Java Connector (SAP JCo) and configure all the connection details required. As soon as you finish the configuration and test the connection, you will get the error message #1:
There is another connectivity issue (error #2) you might face, after completing the steps, in this blog article you will learn how fix both of them:
com.businessobjects.datafederator.sourcedefiner.connection.ConnectionExce ption: com.sap.connectivity.cs.java.drivers.DBDError: [Data Federator Driver] [Server] [Connector 'SAP Testing'] No RFC destination was defined for the server 'SIA.AdaptiveProcessingServer2'
Step 2: Creating a RFC destination in SAP Logon App:
By using the transaction SE37 and using the function module "RSDRI_DF_CONFIGURE" you will create the RFC destination.
This is the crucial part of this configuration, the RFC destination must be created specifically in the same way SAP recommends in this link:
Take a look at the section "Steps for Manual Configuration" and complete your RFC destination as follows:
- Make sure you set an empty value for the I_ONLY_CHECK parameter
- Make sure you set an empty value for the I_REMOVE_CONFIGURATION
- Complete I_RFC_DESTINATION as requested. To give an example if your SAP BusinessObjects server is called MYBOSRV the value for I_RFC_DESTINATION must be DF_JCO_MYBOSRV_1
After doing this, error #2 will be gone.
Step 3: Configuring Data Federation Administration tool properly
In Data Federation Administration Tool, under "Connector Configuration" you will find the Relational Connection created in IDT. Create a configuration for the connection mentioned accordingly with the content in the SAP Logon transaction SM59 - TCP/IP Connections (where you will find the DF_JCO_MYBOSRV_1 connection created in Step2).
You must make sure that these three parameters have the correct values as per your DF_JCO_MYBOSRV_1 in SM59:
- gatewayServiceName: same value as in SM59
- gatewayHostname: same value as in SM59
- programIDMapping: concatenate your SIA name + Adaptive Processing Server name = DF_JCO_MYBOSRV_1
Step 4: Modifying "reginfo.dat" file in BW system:
By using SAP Logon transaction SMGW you will have access to the "reginfo.dat" file in which we will perform a quick change. Navigate to Goto - Expert Functions - External Security - Display (Reg. Info).
As per SAP note 1806694, you will add one line in this file. The easiest way to perform that for me was to search in the SAP BW server for the file "reginfo.dat" and edit it manually. After this, go back to SMGW transaction in SAP Logon and go to Goto - Expert Functions - External Security - Reread (global). You should see the line added in SAP Logon when you display the file again. The line to be added is: P TP=* HOST=<server as per ProgramIDMapping> CANCEL=* ACCESS=*
Where <server as per ProgramIDMapping> means your BO server where SAP Explorer is installed.
If you have completed all the steps in this blog and you try to test your relational connection in IDT, it will work!
So now, you are able to test the relational connection in IDT as well as able to create a Data Foundation and Business Layer on top of it. After publishing the Relational Universe into your repository, you will find the universe avaiable in SAP BusinessObjects Explorer and you will be able to create an Information Space and index it.
Don’t you just hate it when you see a report with DIV/0 errors? Thankfully, Web Intelligence provides us with a function to handle these types of errors, allowing us to provide a more meaningful message.
Let's look at IsError:
- IsError() is a function which will return a boolean value.
- A 1 is returned is the formula being evaluated is in error. A 0 if there is no error.
- This function is particularly useful when embedded into an IF statement.
Let's look into this with an example:
- I’m going to create a variable in a WebI document called formula.
The value for this variable is =6/0.
If we use this variable in the report, we can see something that the picture below:
2. So now is when we can use the function mentioned above.
I create a new variable called IsError to manage this error.
As you can’t divide by zero, a DIV/0 error has occurred and a 1 is returned by the IsError statement.
The IF statement then presents one of two outcomes depending on this boolean value.
In the case of an error, the user is presented with some meaningful text:
“The denominator object is 0”
So if we use this new variable in the report, we can see something that the picture below:
As you can see, using the new variable IsError you can’t see the error message that we got with the variable formula.
3. Now, I’m going to change de value for the variable called formula as below:
The result that we get is the same for both variables:
As conclusion, it is good practice to include error handling into your reports, particularly when you are performing any type of division task.
For the user, it’s much better indicate an explanation of the displayed error that show only the message that WebI provides.
Wrapping an If IsError check around your formula ensures that these type of errors are handled in the correct manner and should help the user identify how the problem can be fixed.
You can manage other kind of error doing the same process.
Knowing your business and making smarter decision in order to keep up with the everyday risks involved in running your business is critical in order to survive in today’s global economy. Making effective decisions requires information. This information must be accurate and updated, and in the right level of detail that you need to be able to move forward at optimal speed. Key business analytics is what allows you to draw information from the data you collect in the different parts of your business. When you understand exactly what is driving your business, where new opportunities come from, and where mistakes were made, you can be proactive to maximize existing revenues and reveal areas for expansion. Better decisions can be made when you have more visibility into vital insight coming from your own company.
SAP Business Intelligence (BI) Solutions provides a window into your company. A dashboard for example is a single, reliable, and real-time overview of your company. It offers you quick insight, in visual appealing formats that are easy to understand. You also have "What-If" tests that let you measure the business impact of a particular change. This can also be made available on mobile devices, so you can make informed decisions on-the-move. When you have information you can trust, you are able to act rapidly and stay ahead of the game.
With years of expertise in BI, Clariba has helped several companies to draw insight from their data. For example, Vodafone Turkey´s Marketing department sought our help to provide the Customer Value Management Team with a dynamic and user-friendly visualization and analysis tool for marketing campaigns. With the central dashboard we delivered, the marketing team was able to analyze existing campaigns and design outlines for new ones based on key success factors.
You can learn more about SAP BI solutions here, and you can also watch SAP BI Solutions videos on YouTube. Want to unlock this information on what drives your business forward? Contact us on info@clariba or leave a comment below, and discover how SAP BI Solutions can help you achieve it.
To continue with my Social Media series (read the previous blog here BI and Social Media – A Powerful Combination Part 1: Google Analytics), today I would like to talk about the biggest social network of them all: Facebook. In this blog post, I will explain different alternatives I have recently researched to extract and use information from Facebook to perform social media analytics with SAP BusinessObjects’ report and dashboard tools. In terms of the amount of useful information we can extract to perform analytics, I personally think that Twitter can be as good or even better than Facebook, however, it has around 400 million less users. Facebook still stands as the social network with the most users around the world - 901million at this moment - making it a mandatory reference in terms of social media analytics.
Before we start talking about technical details, the first thing you should understand is that Facebook is strongly focused on user experience, entertainment applications, content sharing, among others. Therefore, the user activity is more dispersed and variable as opposed to the real-time orderly fashion that Twitter gives us, which is so useful when building trends and chronological analysis. Hence, be sure of what you are looking for, stay focused on your key indicators and make sure you are searching for something that is significant and measurable.
Relevant Facebook APIs for Analytical purposes
The APIs (Application Programming Interface) that Facebook provides are largely directed at the development of applications for social networking and user entertainment. However, there are several APIs that can provide relevant information to establish Key Indicators that can later be used to run reports. As Facebook’s developer page1 states: “ We feel the best API solutions will be holistic cross API solutions.” Among the API’s that you will find most useful (labeled by Facebook as Marketing APIs), I can highlight the Graph API, the Pages API, the Ads API and the Insights API. In any case, I encourage you to take a look at Facebook pages and guides for developers, it will be worth your time:
- Marketing Developer Program: http://developers.facebook.com/preferredmarketingdevelopers/why_build/
- Marketing Developer Resources (with mentions of the APIs above): http://developers.facebook.com/preferredmarketingdevelopers/resources/
- Facebook Marketing Solutions: http://www.facebook.com/marketing
Third party applications to extract data from Facebook
I only found a few third party applications to extract data from Facebook’s API that were comprehensive enough to ensure reliable access to data. Below are some alternatives designed for this requirement:
GA Data Grabber: This application has a module for the Facebook APIs, which costs 500USD a year. As in the case of Google Analytics, it has key benefits such as ease-of-use and flexibility to make queries. It may also be integrated with some tools from SAP BusinessObjects such as WebIntelligence, Data Integrator or Xcelsius dashboards through LiveOffice.2
Custom Application Development: It is the most popular option, as I already mentioned in my previous post about Google Analytics. The Facebook APIs admit access from common programming languages, allowing to record the results of the queries in text files that can be loaded into a database or incorporated directly into various tools of SAP BusinessObjects.
Implementation of a Web Spider: If the information requirements are more focused on the user’s interactions with your client’s Facebook webpage or any of its related Facebook applications, this method may provide complementary information to that which is available in the APIs. The information obtained by the web spider can be stored in files or database for further integration with SAP BusinessObjects tools. Typically, web spiders are developed in a common programming language, although there are some cases where you can buy an application developed by third parties, as the case of Mozenda.3
As I have mentioned in my previous post, in the area of Social Media new applications and trends are appearing at a hectic pace, a lot of changes are expected to happen, so it is just a matter of time until we have more and better options available. I encourage you to stay curious on Social Media analytics and its most popular networks, because right now this is a growing gold mine of information.
If you have any questions or anything to add to help improve this post, please feel free to leave your comments. You may also find interesting the previous post I wrote about SAP BusinessObjects and Google Analytics: http://juancaruiz.com/clariba/bi-and-social-media-a-powerful-combination-part-1-google-analytics/
1 Marketing Developer Program: http://developers.facebook.com/preferredmarketingdevelopers/why_build/
2 GA Data Grabber: http://www.gadatagrabbertool.com/
3 Mozenda: http://www.mozenda.com/
On my previous blog post B-tree vs Bitmap indexes - Indexing Strategy for your Oracle Data Warehouse I answered two questions related to Indexing: Which kind of indexes can we use and on which tables/fields we should use them. As I promised at the end of my blog, now it´s time to answer the third question: what are the consequences of indexing in terms of time (query time, index build time) and storage?
Consequences in terms of time and storage
To tackle this topic I’ll use a test database with a very simplified star schema: 1 fact table for the General Ledger accounts balances and 4 dimensions - the date, the account, the currency and the branch (like in a bank).
To give an idea of the table size, Fact_General_Ledger has 4,5 million rows, Dim_Date 14 000, Dim_Account 3 000, Dim_Branch and Dim_Currency less than 200.
We’ll suppose here that the users could query the data with filter on the date, branch code, currency code, account code, and the 3 levels of the Balance Sheet hierarchy (DIM_ACCOUNT.LVLx_BS) . We assume that the descriptions are not used in filters, but in results only.
Here is the query we will use as a reference:
from fact_general_ledger f
join dim_account a on f.account_key = a.account_key
join dim_date d on f.date_key = d.date_key
join dim_branch b on f.branch_key = b.branch_key
join dim_currency c on f.currency_key = c.currency_key
a.lvl3_bs = 'Deposits With Banks' and
d.date_date = to_date('16/01/2012', 'DD/MM/YYYY') and
b.branch_code = 1 and
c.currency_code = 'QAR' -- I live in Qatar ;-)
So, what are the results in terms of time and storage?
Some of the conclusions we can draw from this table are:
- Using indexes pays off: queries are really faster (about 100 times), whatever the chosen index type is.
- Concerning the query time, the index type doesn’t seem to really matter for tables which are not that big. It would probably change for a fact table with 10 billion rows. There seems however to be an advantage to bitmap indexes and especially bitmap join indexes (have a look at the explanation plan cost column).
- Storage is clearly in favor of bitmap and bitmap join indexes
- Index build time is clearly in favor of b-tree. I’ve not tested the index update time, but the theory says it’s much quicker for b-tree indexes as well.
Ok, I´m convinced to use Indexes. How do I create/maintain one?
The syntax for creating b-tree and bitmap indexes is similar:
Create Bitmap Index Index_Name ON Table_Name(FieldName)
In the case of b-tree indexes, simply remove the word “Bitmap” from the query above.
The syntax for bitmap join indexes is longer but still easy to understand:
create bitmap index ACCOUNT_CODE_BJ
where fact_general_ledger.account_key = dim_account.account_key
Note that during your ETL, you’d better drop/disable your bitmap / bitmap join indexes, and re-create/rebuild them afterwards, rather than update them. It is supposed to be quicker (however I’ve not made any tests).
The difference between drop/re-create and disable/rebuild is that when you disable an index, the definition is kept. So you need a single line to rebuild it rather than many lines for the full creation. However the index build times will be similar.
To drop an index: “drop index INDEX_NAME”
To disable an index: “alter index INDEX_NAME unusable”
To rebuild an index: “alter index INDEX_NAME rebuild”
The conclusion is clear: USE INDEXES! When properly used, they can really boost query response times. Think about using them in your ETL as well: making lookups can be much faster with indexes.
If you’d like to go any further I can only recommend that you read the Oracle Data Warehousing Guide. To get it just look for it on the internet (and don’t forget to specify the version of your database – 10.2, 11.1, 11.2, etc.). It’s a quite interesting and complete document.
Some time ago we’ve seen how to create and maintain Oracle materialized views in order to improve query performance. But while materialized views are a valuable part of our toolbox, they definitely shouldn’t be our first attempt at improving a query performance. In this post we’re going to talk about something you’ve already heard about and used, but we will take it to the next level: indexes.
Why should you use indexes? Because without them you have to perform a full read on each table. Just think about a phone book: it is indexed by name, so if I ask you to find all the phone numbers of people whose name is Larrouturou, you can do that in less than a minute. However if I ask you to find all the people who have a phone number starting with 66903, you won’t have any choice but reading the whole phone book. I hope you don’t have anything else planned for the next two months or so.
It’s the same thing with database tables: if you look for something in a non-indexed multi-million rows fact table, the corresponding query will take a lot of time (and the typical end user doesn’t like to sit 5 minutes in front of his computer waiting for a report). If you had used indexes, you could have found your result in less than 5 (or 1, or 0.1) seconds.
I’ll answer the following three questions: Which kind of indexes can we use? On which tables/fields shall we use them? What are the consequences in terms of time (query time, index build time) and storage?
Which Kind Of Indexes Can We Use?
Oracle has a lot of index types available (IOT, Cluster, etc.), but I’ll only speak about the three main ones used in data warehouses.
B-tree indexes are mostly used on unique or near-unique columns. They keep a good performance during update/insert/delete operations, and therefore are well adapted to operational environments using third normal form schemas. But they are less frequent in data warehouses, where columns often have a low cardinality. Note that B-tree is the default index type – if you have created an index without specifying anything, then it’s a B-tree index.
Bitmap indexes are best used on low-cardinality columns, and can then offer significant savings in terms of space as well as very good query performance. They are most effective on queries that contain multiple conditions in the WHERE clause.
Note that bitmap indexes are particularly slow to update.
Bitmap Join Indexes
A bitmap join index is a bitmap index for the join between tables (2 or more). It stores the result of the joins, and therefore can offer great performances on pre-defined joins. It is specially adapted to star schema environments.
On Which Tables/Fields Shall We Use Which Indexes?
Can we put indexes everywhere? No. Indexes come with costs (creation time, update time, storage) and should be created only when necessary.
Remember also that the goal is to avoid full table reads – if the table is small, then the Oracle optimizer will decide to read the whole table anyway. So we don’t need to create indexes on small tables. I can already hear you asking: “What is a small table?” A million-row table definitely is not small. A 50-row table definitely is small. A 4532-row table? I´m not sure. Lets run some tests and find out.
Before deciding about where we shall use indexes, let’s analyze our typical star schema with one fact table and multiple dimensions.
Let’s start by looking at the cardinality column. We have one case of uniqueness: the primary keys of the dimension tables. In that case, you may want to use a b-tree index to enforce the uniqueness. However, if you consider that the ETL preparing the dimension tables already made sure that dimension keys are unique, you may skip this index (it’s all about your ETL and how much your trust it).
We then have a case of high cardinality: the measures in the fact table. One of the main questions to ask when deciding whether or not to apply an index is: “Is anyone going to search a specific value in this column?” In this example I´ve developed I assume that no one is interested in knowing which account has a value of 43453.12. So no need for an index here.
What about the attributes in the dimension? The answer is “it depends”. Are the users going to do searches on column X? Then you want an index. You’ll choose the type based on the cardinality: bitmap index for low cardinality, b-tree for high cardinality.
Concerning the dimension keys in the fact table, is anyone going to perform a search on them? Not directly (no filters by dimension keys!) but indirectly, yes. Every query which joins a fact table with one or more dimension tables looks for specific dimension keys in the fact table. We have got two options to handle that: putting a bitmap key on every column, or using bitmap join keys.
Are indexes that effective? And what about the storage needed? And the time needed for constructing/ refreshing the indexes?
We will talk about that next week on the second part of my post.
Descubra como las aplicaciones de análisis de negocio de SAP le pueden ayudar a acceder a información en la que basar sus decisiones con seguridad y actuar para mejorar el rendimiento de su empresa.
Atienda a nuestro evento el 4 de Julio de 10:30 a 12:30 y descubra lo que su empresa puede lograr con las soluciones SAP.
En el SAP Roadshow Innova Pyme 2012 usted tendrá la oportunidad de hablar con los consultores de com&Geinsa y Clariba, partners de SAP con años de experiencia en soluciones de gestión de la información. Ellos le proporcionarán información sobre cualquier consulta que formule, o concretaremos el poderle ver con una atención más personalizado.
- Cafes y pastas de bienvenida
- Presentación de SAP BI para BusinessOne
- Demostración del Cuadro de Mandos para Finanzas y Ventas
- Preguntas y respuestas
- Agradecimientos y Despedida
Un cordial saludo,
Marc Haberland Director General Clariba
Acerca de Clariba
Clariba ofrece soluciones innovadoras y fiables en business intelligence, proporcionando a clientes en más de 15 países la visión que necesitan para mejorar su rendimiento empresarial. Nuestros consultores son certificados y expertos en la planificación, desarrollo y instalación de SAP BusinessObjects. www.clariba.com
Acerca de com&Geinsa
com&Geinsa lleva más de 30 años aportando soluciones para rentabilizar al máximo los sistemas de información de su empresa, sin perder de vista la evolución de la tecnologías como movilidad y cloud. En com&Geinsa somos expertos en hacer de SAP una solución única para su empresa. www.comgeinsa.com
Clariba recently carried out a survey about information management with participants at an event. The results were quite interesting since we had most participants mentioning the same pain points; therefore we have decided to share them in this article. To carry out the survey, we used simple statements that people could identify themselves with - divided into two areas: Business and Technical. We then asked participants to fill a one-page questionnaire, asking them to check all the pain points they were facing (they could check multiple pain points).
Business Pain Points
When it came to the business side, 45% of respondents affirmed that their boss constantly pushes them for more information. This is a very sensitive topic. We all know that companies can´t afford to make decisions that aren´t based on facts. Therefore if you are the responsible for gathering these facts, you will be under constant pressure. This would be something simple to deal with if it was easy to gather the information you need. However, when we see that 35% of participants complained they are spending too much time consolidating information and 30% did not manage to get all the information they need, we understand why this becomes a big issue. Business intelligence platforms should not only provide you with all the insight you need, but should make it easily available. It can be a daunting task when you have to wait for IT to return queries for a business question you have to answer or when you have to compare 3 different sources of data to find out the true value before a meeting taking place 1 hour from now.
Another common dissatisfaction was around data obsolescence. 40% said their BI systems only provides historical data and they can´t see how the company is performing right now. This can be caused by not having a proper data warehouse in place, or having one that is not refreshed as often as it should or having poorly designed ETL processes. Whatever the reason is, it can lead to poor decision making. When business moves at the speed of light (in a call center for example), to rely on old information can be misleading.
Another issue related to data obsolescence appeared when 40% said they like Microsoft Excel as a tool to manipulate information, but they are not getting fresh data. If you have to prepare presentations or send emails that include Microsoft Excel data, and you have to manually update them not only will you spend a lot of your time on it, but you are also subject to making mistakes. Business intelligence tools must allow users to optimize their time, while reducing manual labor thus minimizing mistakes.
Technical Pain Points
On the technical side of our survey, the most common complaint was around workload related to information management. 55% of participants stated that their department is completely overloaded. This is also connected to another popular complaint: 45% of respondents state that they have system users calling them all the time. If users don´t have the autonomy to create their own reports and need IT for every query they have to run, the department will become a bottleneck. That´s why offering business users easy-to-use tools that give them the freedom they need to draw insight from corporate data on their own is so important, as it speeds up the information sharing process.
Another frequently mentioned pain point on the technical side is that 40% respondents are not getting the support they need for their BI system. When a problem happens in the BI platform, the impact on a company can be huge, as people make their decisions without trustworthy data to guide them. Therefore, the right support is vital to the correct function of a business, both in SAP licenses and on the BI solutions implemented.
Finally, 35% of respondents don't have the visibility into how their BI systems are being used. When IT has no idea how the system is being used, they cannot optimize the server and the licensing, they can´t understand their users and they can´t be proactive, forecasting when changes will be needed. They need to be able to understand what is going on in their BI applications and ensure BI environments are performing appropriately.
How to Solve These Issues?
Whether you use Microsoft Excel, a legacy BI system or an SAP BI solution, Clariba can help you solve data management issues that you face on a daily basis. Clariba´s solutions, which are based on SAP BusinessObjects, the leading BI platform in the market today, can help business users and IT departments to resolve their concerns.
With the full range of SAP BusinessObjects BI modules we are able to provide companies with the solutions they need. We are also capable of developing optimal ETL processes and Data Warehousing with our Data Management services to make sure you have high quality data analysis.
Furthermore, we can ensure your systems are performing to their maximum capacity. Our Support Center has been certified by SAP to provide support on SAP BusinessObjects licenses and also on BI systems and implementation. If what you are looking for is to monitor your BI system, we have recently developed PerformanceShield, a solution that ensures an effective management of your Business Intelligence (BI) system, which is critical to guarantee your organization is optimizing its investments on BI.
Not sure where to begin? The first step is a 360 BI Assessment, based on best practices and an in-depth understanding of your BI platform, tools, content and licenses. If your current BI system has any gaps or challenges, SAP BusinessObjects and Clariba will recommend steps for making it more secure, stable or efficient for all your users and audiences.
If you feel any of the pain points mentioned, or any of the other ones that appear in the graphs, contact Clariba and find out how you can make the best of your investments in business intelligence.
The set up process for SAP BusinessObjects BI 4.0 when there is a firewall between the servers and the users it´s a bit different than the one from the previous SAP BusinessObjects releases. In this blog post I will walk you through this process.
In order to allow the users to work with the client tools we need to follow the steps below:
Open the ports 8080 on the firewall (or HTTP port you are using) and the range from 6400 to 6410 (you would normally only use the ports 8080,6400,6401,6402, 6403 and 6410, but it is recommended to open the full range, just in case). The ports need to be opened “bi-directionally”. In the advent that you are not using the default port (6410) for the Server Intelligence Agent (SIA) you would also need to open the port you are in fact using.
1- Login into the Central Management Console (CMC) as Administrator, then go to Servers - CMS - Properties. On Common Settings deselect the check-box Auto-Assign and on the text-box assign the port 6401. Save and close.
2- Go to Properties of the Input File Repository server. On Common Settings deselect the checkbox Auto-Assign and on the textbox assign the port 6402.
3- Go to the properties of the Output File Repository server. On Common Settings unselect the checkbox Auto-Assign and on the textbox assign the port 6403.
4- Finally, go to the properties of the Adaptive Processing server. On Common Settings unselect the checkbox Auto-Assign and on the textbox assign the port 6404.
5- The last step is to restart the SIA
You are done! After running this process the SAP BI 4.0 client tools should work for all the users in your network.
The risks of not hard coding the ports to certain servers as observed are: not being able to log in the BI platform from client or web access, not being able to manage content – objects such documents or universes and instances - and not being able to use the multi-Source engine for the new generation universes from the Information Design Tool.
If you have any doubts or would also like to share a tip on this topic please leave a comment below.